How Blamer works

Three surfaces. One Attribution Engine. Every scan answers three questions: What is wrong? How severe? Who wrote it — human or which AI tool?

The pipeline

SCAN -> DETECT -> ATTRIBUTE -> REPORT

  URL / PR / local code
    -> Headless browser captures DOM + network + console
      -> Detection engines find issues across 8 categories
        -> Attribution engine classifies: human or AI tool X
          -> Report with per-issue blame data + per-tool quality profile
    

Three delivery surfaces

1. URL Scanner

Paste any public URL. Blamer launches a headless Chromium, captures the DOM + network + console, runs eight detection engines, then attributes each finding via fingerprinting (no repo access required).

  • Total scan time: ~11 seconds
  • Confidence range without repo access: 0.50–0.85
  • Zero friction: no GitHub OAuth, no install
  • Use case: Compliance audits, vendor assessments, CISO production sweeps

Try a live URL scan

2. GitHub App

2-click install on any GitHub organization. On every PR Blamer scans changed files, runs git blame + Co-Authored-By header analysis, and posts an attribution table as a PR comment. Branch protection optional.

  • Confidence range with repo access: 0.80–0.99
  • Free for public repos (unlimited)
  • SARIF export for GitHub Code Scanning
  • Use case: PR-level review prioritization, CI/CD quality gates

3. CLI Tool

Run locally before commit, or in any CI/CD pipeline:

npx @blamer/cli scan ./src
npx @blamer/cli scan https://example.com
  • Pre-commit hook or pipeline step
  • Configurable severity thresholds per AI source
  • SARIF, JSON, and human-readable terminal output
  • Use case: DevSecOps quality gates with attribution-aware thresholds

Eight detection categories

Category Detects Engine Severity
Data leaksAPI keys, PII exposure, secrets in DOMTruffleHog patterns + customCritical
PerformanceLCP/CLS/INP regressions, memory leaks, N+1Lighthouse + customHigh
SecurityXSS, SQLi, SSRF, insecure endpointsCustom rules + OWASP patternsCritical
AccessibilityWCAG 2.1 AA violationsaxe-coreHigh
AI complianceEU AI Act Art. 50 transparency, AI-content markingEU AI Act rules + customHigh
Regulatory / ISOISO 27001, MDR, PCI DSS, HIPAA, SOC 2 gapsFramework mappingsMedium
SEOSchema.org, heading hierarchy, meta tagsCustom rulesMedium
SustainabilityWSG violations, carbon footprint, green hostingWSG rules + customMedium

Attribution engine (Patent G)

Mode Available when Method Confidence
Git-blame-based Repo access (GitHub App, CLI) Commit metadata, Co-Authored-By, IDE markers, author patterns 0.80–0.99
Fingerprint-based URL scan (no repo) AI naming patterns, code structure, comment style, framework usage 0.50–0.85

Supported AI tool profiles: Copilot, Cursor, Claude Code, Windsurf, Devin, CodeWhisperer, Tabnine. Extensible via profile plugins.

Validation evidence

Patent G's multi-signal AI code attribution methodology was cross-referenced against 6,439,303 code samples spanning 64 distinct AI models (Llama, Qwen, DeepSeek, Gemini, Phi, GPT-4, Claude, IBM Granite, etc.) across 13 programming languages and 9 published datasets (PoC v3.0, March 2026). End-to-end Blamer classifier accuracy on production customer code is pending Patent G PoC v4.0 (industrial-scale, 100M+ samples).

Three confirmed claims

  1. Multi-signal detection necessary. Single-feature classifiers (e.g. code length alone) are insufficient — Cliff's δ 0.07–0.28 (negligible-to-small effect size across 6M+ samples). Patent G's multi-signal architecture is necessary, not optional.
  2. Per-model fingerprinting feasible. Distinct line/character/stdev signatures per LLM (e.g. o3-mini avg 169.9 lines vs. llama3.3 avg 83.2 — a 2× difference). Enables per-tool quality scoring on customer codebases.
  3. AI adoption exponential. DevGPT corpus shows 145% growth in 77 days (Spearman rs = 0.98, p < 0.001); power users dominate (Gini 0.68 — top 7% of authors produce 41% of AI commits).

Source: PATENT_G_POC_V3_REPORT.md. All 7/7 Patent G claims CONFIRMED with STRONG-to-VERY-STRONG evidence at scale.

See pricing →