Back to Qwen Code

CodeScope Q&A

.qwen/skills/codegraph/SKILL.md

0.15.1129.0 KB
Original Source

CodeScope Q&A

CodeScope indexes source code into a two-layer knowledge graph — structure (functions, calls, imports, classes, modules) and evolution (commits, file changes, function modifications) — plus semantic embeddings for every function. Supports Python, JavaScript/TypeScript, C, and Java (including Hadoop-scale repositories with 8K+ files). This combination enables analyses that grep, LSP, or pure vector search cannot do alone. It can also fetch GitHub issues and trace bugs to code, and review open PRs — scoring per-PR risk, detecting cross-PR conflicts, identifying auto-merge candidates, and applying GitHub labels.

When to Use This Skill

  • User asks about call chains, callers, callees, or dependencies
  • User wants to find dead code, hotspots, or architectural layers
  • User asks about code history, who changed what, or why something was modified
  • User wants to find semantically similar functions across a codebase
  • User wants a full architecture analysis or report
  • User asks about module coupling, circular dependencies, or bridge functions
  • User wants to index or analyze a Java project (Maven, Gradle, plain Java)
  • User wants to analyze GitHub issues or bug reports to find root causes
  • User asks "why does this project have so many bugs" or "what code is most buggy"
  • User wants to trace a bug report to the most relevant code locations
  • User asks about class relationships, ownership, composition, or wants a class diagram / UML
  • User wants to understand which classes own or depend on other classes
  • User wants to review PRs, assess PR risk, or prioritize PR reviews
  • User asks about cross-PR conflicts or which PRs can be merged independently
  • User wants to find auto-merge candidates or generate a PR review report
  • User asks about the blast radius or impact scope of a PR
  • User wants to apply labels to PRs from analysis results
  • User wants to explore PR-specific follow-up questions for a given PR
  • A .codegraph directory (or similar index) exists in the workspace

Getting Started

Installation

bash
pip install codegraph-ai

Environment Variables (optional)

bash
# Create Python virtural environment
python -m venv .venv

source .venv/bin/activate

# Point to a pre-built database (skip indexing)
export CODESCOPE_DB_DIR="/path/to/.linux_db"

# Offline mode for HuggingFace models
export HF_HUB_OFFLINE="1"

# Fallback when HuggingFace is unreachable (e.g., network issues in China)
# Use HF mirror or ModelScope for sentence-transformers models:
export HF_ENDPOINT="https://hf-mirror.com"
# https://www.modelscope.cn/models/sentence-transformers/all-MiniLM-L6-v2

Check Index Status

bash
codegraph status --db $CODESCOPE_DB_DIR

If no index exists, create one:

bash
codegraph init --repo . --lang auto --commits 500

Supported languages: python, c, javascript, typescript, java, or auto (auto-detects from file extensions).

The --commits flag ingests git history (for evolution queries). Without it, only structural analysis is available. Add --backfill-limit 200 to also compute function-level MODIFIES edges (slower but enables change_attribution and co_change).

To add git history to an existing index (without re-indexing structure):

bash
codegraph ingest --repo . --db $CODESCOPE_DB_DIR --commits 500
codegraph ingest --repo . --db $CODESCOPE_DB_DIR --backfill-limit 200   # add MODIFIES edges only

Two Interfaces: CLI vs Python

Use the CLI for status and reports:

bash
codegraph status --db $CODESCOPE_DB_DIR
codegraph analyze --db $CODESCOPE_DB_DIR --output report.md

Use the Python API for queries and custom analyses:

python
import os
os.environ['HF_HUB_OFFLINE'] = '1'  # required

from codegraph.core import CodeScope
cs = CodeScope(os.environ['CODESCOPE_DB_DIR'])

# Cypher query
rows = list(cs.conn.execute('''
    MATCH (caller:Function)-[:CALLS]->(f:Function {name: "free_irq"})
    RETURN caller.name, caller.file_path LIMIT 10
'''))
for r in rows:
    print(r)

cs.close()  # always close when done

The Python API is more powerful — it gives you raw Cypher access and lets you chain queries.

Core Python API

Raw Queries

These are the building blocks for any custom analysis:

MethodWhat it does
cs.conn.execute(cypher)Run any Cypher query against the graph — returns list of tuples
cs.vector_only_search(query, topk=10)Semantic search over all function embeddings — returns [{id, score}]
cs.summary()Print a human-readable overview of the indexed codebase

Structural Analysis

MethodWhat it does
cs.impact(func_name, change_desc, max_hops=3)Find callers up to N hops, ranked by semantic relevance to the change
cs.hotspots(topk=10)Rank functions by structural risk (fan-in × fan-out)
cs.dead_code()Find functions with zero callers (excluding entry points)
cs.circular_deps()Detect circular import chains at file level
cs.module_coupling(topk=10)Find cross-module coupling pairs with call counts
cs.bridge_functions(topk=30)Find functions called from the most distinct modules
cs.layer_discovery(topk=30)Auto-discover infrastructure / mid / consumer layers
cs.stability_analysis(topk=50)Correlate fan-in with modification frequency
cs.class_hierarchy(class_name=None)Return inheritance tree for a class (or all classes)

Class Dependency Relationships (UML-Style)

CodeScope extracts three UML relationship types from class fields and type annotations during indexing:

RelationshipUML symbolMeaningHow detected
COMPOSES*-- filled diamondStrong ownership — field always holds an instanceNon-optional field assigned a constructed object
AGGREGATESo-- open diamondOptional/weak reference — may be NoneOptional[X], X | None, or assigned None
INHERITS<|-- hollow arrowSubclass extends parentclass A(B)
python
# Get all composition relationships (A strongly owns B)
list(cs.conn.execute('MATCH (c1:Class)-[:COMPOSES]->(c2:Class) RETURN c1.name, c2.name'))

# Get all aggregation relationships (A optionally holds B)
list(cs.conn.execute('MATCH (c1:Class)-[:AGGREGATES]->(c2:Class) RETURN c1.name, c2.name'))

# How many objects does a class directly own?
list(cs.conn.execute(
    'MATCH (c:Class {name: "Llama"})-[:COMPOSES]->(t:Class) RETURN t.name'
))

# Full dependency graph for a class (composition + aggregation + inheritance)
list(cs.conn.execute(
    'MATCH (c:Class {name: "GPUModelRunner"})-[r:COMPOSES|AGGREGATES]->(t:Class) '
    'RETURN type(r), t.name'
))

Generating a Mermaid class diagram:

python
inherits  = list(cs.conn.execute('MATCH (c1:Class)-[:INHERITS]->(c2:Class) RETURN c1.name, c2.name'))
composes  = list(cs.conn.execute('MATCH (c1:Class)-[:COMPOSES]->(c2:Class) RETURN c1.name, c2.name'))
aggregates = list(cs.conn.execute('MATCH (c1:Class)-[:AGGREGATES]->(c2:Class) RETURN c1.name, c2.name'))

print('classDiagram')
for src, tgt in inherits:   print(f'    {tgt} <|-- {src}')   # parent <|-- child
for src, tgt in composes:   print(f'    {src} *-- {tgt}')    # owner *-- owned
for src, tgt in aggregates: print(f'    {src} o-- {tgt}')    # holder o-- optional

Scale reference:

ProjectClassesINHERITSCOMPOSESAGGREGATESIndex time
llama-cpp-python1281884~2s
vllm4,0022,1853,217149~50s
MethodWhat it does
cs.similar(function, scope, topk=10)Find functions similar to a given function within a module scope
cs.cross_locate(query, topk=10)Find semantically related functions, then reveal call-chain connections
cs.semantic_cross_pollination(query, topk=15)Find similar functions across distant subsystems

Evolution (requires --commits during init)

MethodWhat it does
cs.change_attribution(func_name, file_path=None, limit=20)Which commits modified a function? (requires backfill)
cs.co_change(func_name, file_path=None, min_commits=2, topk=10)Functions that are always modified together
cs.intent_search(query, topk=10)Find commits matching a natural-language intent
cs.commit_modularity(topk=20)Score commits by how many modules they touch
cs.hot_cold_map(topk=30)Module modification density

Report Generation

python
from codegraph.analyzer import generate_report
report = generate_report(cs)  # full architecture analysis as markdown

Or via CLI:

bash
codegraph analyze --output reports/analysis.md

The report covers: overview stats, subsystem distribution, top modules, architectural layers (with Mermaid diagrams), bridge functions, fan-in/fan-out hotspots, cross-module coupling, evolution hotspots, and dead code density.

Java Support

CodeScope includes a full Java adapter that handles enterprise-scale repositories like Apache Hadoop (~8K files, ~97K functions indexed in ~3.5 minutes).

What Gets Indexed

ElementGraph Node/EdgeNotes
ClassesClass nodeIncludes generics, annotations
InterfacesClass nodeextendsINHERITS edge
EnumsClass nodeEnum methods extracted
MethodsFunction nodeFull generic signatures, JavaDoc
ConstructorsFunction node (name=<init>)Including super() calls
Method callsCALLS edgeReceiver context preserved (obj.method())
new expressionsCALLS edge to ClassName.<init>Constructor invocations
ImportsIMPORTS edge (file→file)Single, wildcard, static
Inner classesClass node (name=Outer.Inner)Prefixed with outer class
InheritanceINHERITS edgeextends + implements

Indexing a Java Project

bash
codegraph init --repo /path/to/java-project --lang java --commits 500

Or with auto-detection (auto-detects .java files):

bash
codegraph init --repo /path/to/java-project --lang auto

Java-Specific Exclusions

By default, these directories are excluded when indexing Java projects: target/, build/, .gradle/, .idea/, .settings/, bin/, out/, test/, tests/, src/test/.

Java Query Examples

python
# Find all classes that extend a specific class
list(cs.conn.execute("""
    MATCH (c:Class)-[:INHERITS]->(p:Class {name: 'FileSystem'})
    RETURN c.name, c.file_path
"""))

# Find all methods in a specific class
list(cs.conn.execute("""
    MATCH (c:Class {name: 'DefaultParser'})-[:HAS_METHOD]->(f:Function)
    RETURN f.name, f.signature
"""))

# Find constructor call chains
list(cs.conn.execute("""
    MATCH (f:Function)-[:CALLS]->(init:Function {name: '<init>'})
    WHERE init.class_name = 'Configuration'
    RETURN f.name, f.file_path LIMIT 10
"""))

Bug Root Cause Analysis

CodeScope can fetch GitHub issues and map them to code using the graph + vector infrastructure. This is the core workflow for answering questions like "why does this project have so many bugs?" or "where in the code does this bug come from?"

Prerequisites

  • A code graph must already be indexed for the target repository
  • gh CLI must be installed and authenticated (gh auth login)

Bug Analysis API

Single Issue Analysis

python
# Analyze a specific GitHub issue against the indexed code graph
result = cs.analyze_issue("owner", "repo", 1234, topk=10)
print(result.format_report())

This:

  1. Fetches the issue from GitHub (or loads from cache)
  2. Parses file paths, function names, and stack traces from the issue body
  3. Matches extracted paths to File nodes in the graph
  4. Uses semantic search (cross_locate) to find related code
  5. Traces callers of mentioned functions via impact()
  6. Ranks and returns root cause candidates with explanation

Batch Bug Analysis

python
# Analyze top-k bug issues and get aggregated hotspot data
results = cs.analyze_top_bugs("owner", "repo", k=10, label="bug")
for r in results:
    print(f"#{r.issue.number}: {r.issue.title}")
    for c in r.candidates[:3]:
        print(f"  {c.function_name} ({c.file_path}) score={c.score:.2f}")

CLI Commands

bash
# Fetch and parse a single issue (no graph needed)
codegraph fetch-issue owner repo 1234

# Fetch top-k bugs from a repo
codegraph fetch-bugs owner repo --top 10 --label bug

# Analyze a single bug against the code graph
codegraph analyze-bug owner repo 1234 --db .codegraph --topk 10

# Batch analyze top bugs
codegraph analyze-bugs owner repo --db .codegraph --top 10 --label bug

Lower-Level Components

For custom analysis pipelines, the components can be used individually:

python
from codegraph.issue_fetcher import fetch_and_parse_issue
from codegraph.bug_locator import (
    resolve_paths_to_files,
    find_semantic_matches,
    trace_callers,
    rank_root_causes,
    analyze_bug,
)

# Fetch and parse (with caching)
issue = fetch_and_parse_issue("owner", "repo", 1234)
print(issue.extracted_paths)   # file paths found in body
print(issue.extracted_funcs)   # function names from stack traces
print(issue.linked_commits)    # merge commit SHAs from linked PRs

# Match paths to graph nodes
path_matches = resolve_paths_to_files(cs, issue.extracted_paths)

# Semantic search using issue description
semantic_matches = find_semantic_matches(cs, f"{issue.title}\n{issue.body}")

# Trace callers of mentioned functions
caller_traces = trace_callers(cs, issue.extracted_funcs, max_hops=2)

# Combine into ranked candidates
candidates = rank_root_causes(path_matches, semantic_matches, caller_traces, issue.extracted_funcs)

Scoring System

Root cause candidates are scored by combining multiple signals:

SignalScoreDescription
Direct mention+1.0Function name appears in issue body/stack trace
File path match+0.8Function is in a file mentioned in the issue
Semantic match+scoreRaw cosine similarity (0.0-1.0) from cross_locate
Caller relationship+0.5/hopsFunction calls a mentioned function (decays with distance)

Issue Cache

Parsed issues are cached at ~/.codegraph/issue_cache/{owner}_{repo}_{number}.json. Cache hits skip the GitHub API call entirely (sub-millisecond). To force a refresh, pass use_cache=False or use --no-cache on CLI.

python
from codegraph.issue_cache import clear_cache
clear_cache(owner="openclaw", repo="openclaw")  # clear specific repo
clear_cache()  # clear all

Stack Trace Parsing

The parser automatically extracts file paths and function names from stack traces in Python, C/C++, JavaScript/Node.js, Go, and Rust formats. It also extracts func_name() references in backticks and inline code.

PR Review and Analysis

CodeScope can analyze open PRs against the indexed code graph to compute structural risk scores, detect cross-PR conflicts, and generate prioritized review reports.

Prerequisites

  • A code graph must already be indexed for the target repository
  • gh CLI must be installed and authenticated (gh auth login)
  • GITHUB_TOKEN environment variable recommended to avoid rate limiting

Unified Pipeline (CLI)

Two subcommands: prepare (analyze + write to DB) and label (apply GitHub labels + comments).

bash
# Phase 1: Analyze PRs, detect conflicts, write to graph DB (full rebuild)
# Pipeline: cross-PR analysis → single-PR risk scoring → report + labels
codegraph pr-review prepare --db .codegraph

# Filter by author during prepare:
codegraph pr-review prepare --db .codegraph --author someone

# Override auto-detected GitHub repo (owner/repo):
codegraph pr-review prepare --db .codegraph --repo owner/repo

# Skip per-PR risk scoring (conflict-only, faster):
codegraph pr-review prepare --db .codegraph --skip-single-pr

# Phase 2: Apply labels and post conflict comments from graph DB
codegraph pr-review label --db .codegraph

# Label with dry-run (preview without API calls):
codegraph pr-review label --db .codegraph --dry-run

Required arg: --db. Local repo path derived from --db parent. GitHub repo auto-detected from git remote get-url origin (or specified via --repo). Optional: --author, --output, --skip-single-pr (prepare); --dry-run (label).

Python API (for agents / scripts)

For programmatic use within the same Python process, use PRReview — a high-level wrapper that manages CodeScope lifecycle automatically.

python
from codegraph.pr_api import PRReview

# Full pipeline in 2 lines
with PRReview(db=".codegraph") as pr:
    pr.prepare()                # fetch PRs → graph DB → scoring → report
    pr.label(dry_run=True)      # preview labels without API calls

# Query after prepare (works across sessions once DB has data)
with PRReview(db=".codegraph") as pr:
    # Conflicts
    pr.conflict_prs_of("100")           # → ["101", "102"]

    # Risk
    pr.risk("100")                      # → {"number": "100", "risk_level": "HIGH", ...}

    # Classification
    pr.auto_merge_candidates()          # → [{"number": "200", ...}, ...]
    pr.conflicting_groups()             # → [["100", "101"], ["103"]]

    # All PRs in DB
    pr.all_prs()                        # → [{"number": "100", ...}, ...]

    # Functions changed by a specific PR (added / modified / deleted)
    import json
    cs = pr._open_cs()
    rows = list(cs.conn.execute(
        f"MATCH (pr:PR {{id: {json.dumps('439')}}})-[c:CHANGES]->(f:Function) "
        f"RETURN c.info AS change_type, f.name, f.file_path "
        f"ORDER BY c.info, f.name"
    ))
    for change_type, name, path in rows:
        print(f"  [{change_type}] {name} ({path})")
    # change_type: 'hunk' (modified), 'new' (added), 'deleted', 'related' (newly calls)

All query methods return structured Python objects — no text parsing required. The CLI and Python API share the same underlying implementation (run_prepare / run_label / graph DB), so you can prepare via CLI and query via Python, or vice versa.

For lower-level components (PRScorer, CrossPRAnalyzer, etc.), see:

python
from codegraph.pr_analysis import GitHubClient, GraphAnalyzer, PRScorer, CrossPRAnalyzer
gh = GitHubClient(repo='owner/repo')
scorer = PRScorer(GraphAnalyzer(cs, repo_dir), repo_dir, gh)
result = scorer.analyze(gh.pr_to_entry(pr), output_dir='/tmp')  # risk_score, risk_level, peak_blast...

cross = CrossPRAnalyzer(cs, repo_dir, gh)
cross.prepare(pr_ids)  # index PR nodes into graph
cross.connected_components()  # {root: [pr_ids]} — detects conflicts
cross.update_pr_labels(assignments)  # persist labels to graph DB

# Load PR results from graph DB (no GitHub API needed)
all_results, components = cross.load_from_graph()

# Build and apply labels from analysis results
from codegraph.pr_labeler import build_label_assignments, apply_labels
assignments = build_label_assignments(all_results, components)
apply_labels(assignments, repo='owner/repo', create_labels=True)

For detailed workflows, Cypher patterns, and CrossPRAnalyzer query dimensions, see pr-analysis.md.

Report Structure (3 sections)

  1. Auto-merge Candidates: LOW risk, no interface/config changes, singleton component
  2. Independent Review: Non-trivial PRs with no cross-PR conflict
  3. Conflicting PR Groups: PRs sharing code/call paths via connected-components (DSU)

Risk levels: CRITICAL (≥12), HIGH (≥7), MEDIUM (≥3), LOW (<3), UNKNOWN (when --skip-single-pr). Key signals: blast_radius (3.0×), no_test_coverage (2.0×), interface_change (2.5×), dead_code (1.5×).

Applying Labels and Conflict Comments

After running codegraph pr-review prepare, run codegraph pr-review label to apply category labels to GitHub PRs and post conflict comments:

bash
# Apply labels and post conflict comments:
codegraph pr-review label --db .codegraph

# Preview without making API calls:
codegraph pr-review label --db .codegraph --dry-run

The label subcommand reads PR labels from the graph DB (pr.label column) — no re-analysis needed. For conflicting PRs (labelled conflicting-group-N), it also posts a comment on the GitHub PR listing shared functions and other conflicting PRs.

Labels are computed during prepare from the analysis results (connected components + risk scores) and persisted to PR nodes in the graph DB (pr.label column, semicolon-delimited).

Label scheme:

CategoryLabelColor
Auto-merge Candidates (Part 1)auto-merge-candidateGreen
Independent Review (Part 2)independent-reviewYellow
Conflicting Group N (Part 3)conflicting-group-NRed/Orange/Blue
Any conflicting PR (Part 3)conflicting-prRed

Follow-up Exploration

PR-specific follow-up questions are automatically included in codegraph explore when PR nodes exist in the graph DB (i.e., after codegraph pr-review prepare). PR exploration is a question template set integrated into explore. To query a specific PR's details (conflicts, changed functions), use the PRReview Python API.

bash
# After pr-review prepare, explore includes PR questions automatically:
codegraph explore --db .codegraph --top 15

# Interactive exploration (including PR follow-up questions):
codegraph explore --db .codegraph

# Focus on PR-specific questions (use reviewer role):
codegraph explore --db .codegraph --role reviewer

# Filter to only architecture questions (exclude PR patterns):
codegraph explore --db .codegraph --type architecture

# Filter to only risk questions:
codegraph explore --db .codegraph --type risk

# Filter to only PR review questions:
codegraph explore --db .codegraph --type pr-review --role reviewer

The --type filter controls which question categories appear:

  • all (default): all categories mixed together
  • architecture: structural design questions (fan-in, coupling, cycles)
  • risk: risk-focused questions (structural risk + PR risk)
  • evolution: git history questions (change attribution, modification patterns)
  • hotspot: frequently modified code questions
  • pr-review: PR-specific questions (impact, conflicts, test coverage)

When --type pr-review is specified, only PR-related questions are shown.

How to Route Questions

The key decision is: does the user want an exact structural answer, a fuzzy semantic one, or a bug-to-code mapping?

User asks...Best approach
"Who calls free_irq?"Cypher: MATCH (c:Function)-[:CALLS]->(f:Function {name: 'free_irq'}) RETURN c.name, c.file_path
"Find functions related to memory allocation"cs.vector_only_search("memory allocation") or cs.cross_locate("memory allocation")
"What's the most complex function?"cs.hotspots(topk=1)
"Is there dead code in the networking stack?"cs.dead_code() then filter by file path
"How has schedule() changed recently?"cs.change_attribution("schedule", "kernel/sched/core.c")
"Which modules are tightly coupled?"cs.module_coupling(topk=20)
"Generate a full architecture report"codegraph analyze or generate_report(cs)
"What's the architectural role of mm/?"cs.layer_discovery() then find mm entries
"Which functions act as API boundaries?"cs.bridge_functions(topk=30)
"Find commits about fixing race conditions"cs.intent_search("fix race condition")
"What functions are always changed together with kmalloc?"cs.co_change("kmalloc")
"Why does this project have so many bugs?"cs.analyze_top_bugs("owner", "repo", k=10) then aggregate hotspots
"Analyze issue #1234 from GitHub"cs.analyze_issue("owner", "repo", 1234)
"What code is related to this bug?"cs.analyze_issue(...) or manual cross_locate(bug_description)
"Find the root cause of the crash in issue #42"cs.analyze_issue("owner", "repo", 42)
"Which modules have the most bugs?"cs.analyze_top_bugs(...) then aggregate by file/module
"Index this Java project"codegraph init --repo . --lang java
"What classes extend FileSystem in Hadoop?"Cypher: MATCH (c:Class)-[:INHERITS]->(p:Class {name: 'FileSystem'}) RETURN c.name, c.file_path
"Find all constructors called in this module"Cypher: MATCH (f:Function)-[:CALLS]->(init:Function {name: '<init>'}) WHERE f.file_path CONTAINS 'module' RETURN ...
"Draw a class diagram / show class UML"Query COMPOSES, AGGREGATES, INHERITS edges and render as Mermaid classDiagram
"What does Llama own / compose?"Cypher: MATCH (c:Class {name:'Llama'})-[:COMPOSES]->(t:Class) RETURN t.name
"Which class holds a reference to KVCacheManager?"Cypher: MATCH (c:Class)-[:COMPOSES|AGGREGATES]->(t:Class {name:'KVCacheManager'}) RETURN c.name
"Show all optional dependencies of GPUModelRunner"Cypher: MATCH (c:Class {name:'GPUModelRunner'})-[:AGGREGATES]->(t:Class) RETURN t.name
"Review all open PRs and generate report"codegraph pr-review prepare --db ...
"Which PRs can be auto-merged?"Run pr-review prepare, check Part 1 of report
"Are there conflicting PRs?"Run pr-review prepare, check Part 3 (connected components)
"What's the risk of PR #42?"PRScorer.analyze(entry) for per-PR scoring
"What's the blast radius of this PR?"PRScorer.analyze(entry)result['peak_blast'] and call graph viz
"Which PRs modify the same function?"CrossPRAnalyzer.connected_components() → same-function edge type
"Label PRs with their review category"codegraph pr-review label --db ...
"Post conflict comments on PRs"codegraph pr-review label --db ... (automatic for conflicting PRs)
"Preview labels/comments without applying"codegraph pr-review label --db ... --dry-run
"Explore PR follow-up questions interactively"codegraph explore --db .codegraph (auto-includes PR patterns if prepare was run)
"Query a specific PR's conflicts"PRReview.conflict_prs_of("42") — returns list of conflicting PR numbers
"Query a specific PR's changed functions"Cypher: MATCH (pr:PR {id: '42'})-[c:CHANGES]->(f:Function) RETURN c.info, f.name, f.file_path
"Compare two PRs for overlap"Cypher: MATCH (pr1:PR {id: '42'})-[c1:CHANGES]->(f:Function)<-[c2:CHANGES]-(pr2:PR {id: '43'}) RETURN f.name, f.file_path
"Show only architecture questions"codegraph explore --db .codegraph --type architecture
"Show only PR review questions"codegraph explore --db .codegraph --type pr-review --role reviewer
"Show top PR risk questions"codegraph explore --db .codegraph --top 15 --role reviewer
"Full PR review pipeline: analyze, label, explore"1) codegraph pr-review prepare 2) codegraph pr-review label 3) codegraph explore --db .codegraph

For novel investigations not covered by pre-built methods, compose raw Cypher queries. See patterns.md for templates. For bug analysis patterns, see bug-analysis.md.

Important Filters for Cypher

When writing Cypher queries, these filters prevent misleading results:

  • f.is_historical = 0 — exclude deleted/renamed functions that are still in the graph as historical records
  • f.is_external = 0 (on File nodes) — exclude system headers/library files
  • c.version_tag = 'bf' — only backfilled commits have MODIFIES edges; non-backfilled commits only have TOUCHES (file-level) edges
  • Always use LIMIT — large codebases can return hundreds of thousands of rows

Checking Data Availability

Before running evolution queries, check what's available:

python
# How many commits are indexed?
list(cs.conn.execute("MATCH (c:Commit) RETURN count(c)"))

# How many have MODIFIES edges (backfilled)?
list(cs.conn.execute("MATCH (c:Commit) WHERE c.version_tag = 'bf' RETURN count(c)"))

If no commits exist, evolution methods will return empty results — guide the user to run codegraph ingest first. If commits exist but aren't backfilled, TOUCHES (file-level) queries still work but MODIFIES (function-level) queries won't.

Troubleshooting

ErrorCauseFix
Database lockedCrashed process left neug lockrm <db>/graph.db/neugdb.lock
Can't open lock filezvec LOCK file deletedtouch <db>/vectors/LOCK
Can't lock read-write collectionAnother process holds lockKill the other process
recovery idmap failedStale WAL filesRemove empty .log files from <db>/vectors/idmap.0/
HuggingFace model download failsNetwork/firewall blocks huggingface.coUse HF_ENDPOINT="https://hf-mirror.com" or ModelScope (see Getting Started tip)

The CLI auto-cleans lock issues on startup when possible.

References

  • schema.md — Full graph schema: node types, edge types, properties, Cypher syntax notes
  • patterns.md — Ready-to-use Cypher query templates and composition strategies
  • bug-analysis.md — Bug analysis workflows: single issue, batch analysis, hotspot aggregation, custom pipelines
  • pr-analysis.md — PR analysis workflows: per-PR scoring, cross-PR conflict detection, Cypher patterns, CrossPRAnalyzer usage