18
Tools analyzed
What each code intelligence tool can and can't build, based on architecture. 18 tools scored across 45 criteria. Interactive map, weight sliders, and evidence-based competitor analysis.
18
Tools analyzed
101
roam MCP tools (23 core)
139
roam CLI commands
2026-02-27
Last matrix refresh
Y-axis: analysis depth. X-axis: agent readiness. Click any point for details.
18 tools scored across 45 criteria in 7 categories. Adjust weights below to apply your own priorities.
Drag sliders to reweight categories. Map positions and total scores update in real-time.
This scoring system was designed by the roam project maintainers. While 97% of criteria (44/45) are binary or count-based, three forms of bias exist: criteria selection (we chose what to measure), category weights (we set defaults), assessment (we scored all competitors). The weight sliders, open-source data, and per-criterion visibility are mitigations.
Architecture determines capability ceilings. Here's what each tool can and can't build -- and where roam stands honestly against each.
Static Analysis & Security
The enterprise standard for code quality. Full intra-function dataflow and taint propagation. Research-backed Cognitive Complexity (they defined the spec). Quality Gates (they invented the concept).
GitHub's security analysis engine. Full source-to-sink taint tracking with a custom query language for expressing complex vulnerability patterns. Powers GitHub Code Scanning and Copilot Autofix.
Fast structural pattern matching with a large community rules marketplace. Lightweight by design -- no persistent index, no dependency model. Semgrep Pro adds inter-procedural taint.
Code Intelligence (Our Space)
The closest MCP competitor by tool count. SCIP-based semantic indexing gives precise type-aware cross-references. Designed specifically for AI agent workflows with compound and batch operations.
Enterprise-scale code search and navigation across thousands of repositories. Invented the SCIP protocol for precise code intelligence. Code monitoring and batch changes.
Shares our architectural DNA -- tree-sitter parsing into a NetworkX graph with Louvain community detection. Simpler scope, focused on basic graph queries. 14 MCP tools.
Rust-based graph analysis engine with a Universal AST abstraction. Newer project, focused on fast parsing and multi-language AST uniformity. 20 MCP tools.
Behavioral Analysis
Git history as the primary data source. Temporal coupling, team dynamics, knowledge distribution. Behavioral signals that complement structural analysis. The only tool with deeper team-level analysis than roam.
Different architectures answer different questions. Understanding these trade-offs gives the complete picture beyond point comparisons.
SonarQube has 6,500+ rules because each rule is language-specific. "Don't use eval" exists as separate rules for JavaScript, Python, PHP, Ruby, Perl, and Groovy. "Max function length" is another 30 language-specific copies. ~200 unique concepts × 30 languages = 6,000+ rules.
Tools built on tree-sitter (roam, CodeGraphMCPServer, CodePrism) define rules that work across all supported languages from a single definition. One cross-language rule replaces 27 language-specific rules. 100 well-designed cross-language rules can match the coverage of 3,000+ per-language rules.
Rule-first tools (SonarQube, CodeQL, Semgrep) scan files individually against a catalog of patterns. They excel at: "Is this line vulnerable?" "Does this code match a known bad pattern?" Their strength is depth -- tracking data through variables, branches, and function calls to find exploitable paths.
Graph-first tools (roam, CodeScene) model the codebase as a connected system and compute structural properties. They excel at: "What breaks if I change this?" "Where are the architectural bottlenecks?" "Which modules are too coupled?" Their strength is breadth -- seeing the forest, not just individual trees.
These approaches are complementary: graph analysis finds architectural risks that no amount of line-level scanning can detect. Taint analysis finds security vulnerabilities that no graph algorithm can trace.
Tools evaluated but excluded from the landscape. They overlap with code intelligence in narrow ways but don't build a persistent structural model of your codebase.
Compact comparison aligned to the source tracker.
Based on primary product documentation, public repositories, and direct feature verification. Scores are computed from 45 criteria across 7 categories. 44 are binary or count-based. 1 (documentation quality) is subjective and marked as such.
Snapshot date: February 27, 2026. Vendor claims change quickly; this page is version-pinned and refreshed with explicit evidence notes.