Skip to content
Scannica
v0.1 · pre-launch · invite list open

Audit the web
that AI reads.

Scannica is a desktop auditor with 130+ checks across SEO, AI-readiness, accessibility, security, performance, GDPR, and best practices. The crawl runs on your machine. Nothing is uploaded.

Categories
7
Rules
130+
Uploaded
0B
Scannica · Smart Report
v0.1.0
Smart Report Pages URLs Sitemap Carbon
216 URLs
63
/100
www.example.com · Last crawl 2h ago
−7 quality vs last crawl
3 critical 21 warnings 9 notices
GEO READINESS
9 issues
  • Warning GEO_NO_LLMS_TXT site-wide
  • Warning GEO_NO_AUTHOR_ENTITY 47 pages
  • Warning GEO_LOW_ENTITY_DENSITY 38 pages
  • Notice GEO_NO_FAQ_SCHEMA 12 pages
SECURITY
12 issues
  • Error SEC_CSP_UNSAFE_EVAL /checkout
  • Error SEC_COOKIE_NO_HTTPONLY session.id
  • Warning SEC_HSTS_MISSING site-wide
local · 0 bytes sent
⏎ Open in app
01 · what it catches

Seven categories. 81+ rules. One of them, your auditor isn't running.

Every finding carries a stable rule code, a severity, a description, and where applicable a CWE or WCAG reference. The codes match what the app exports, ships in tickets, and writes into .scannica bundles.

SEO
22
6 err
9 warn
7 notice
  • TITLE_MISSING
  • TITLE_TOO_LONG
  • TITLE_TOO_SHORT
  • +19 more
22 rules
Performance
9
2 err
6 warn
1 notice
  • PERF_LARGE_PAGE
  • PERF_HUGE_PAGE
  • PERF_DOM_SIZE
  • +6 more
9 rules
Accessibility
9
3 err
5 warn
1 notice
  • IMG_ALT_MISSING
  • HEADING_SKIP
  • FORM_LABEL_MISSING
  • +6 more
9 rules
Best Practices
11
0 err
4 warn
7 notice
  • BP_NO_DOCTYPE
  • BP_NO_CHARSET
  • BP_CONSOLE_LOG
  • +8 more
11 rules
Security
15
7 err
6 warn
2 notice
  • SEC_NOT_HTTPS
  • SEC_MIXED_CONTENT_ACTIVE
  • SEC_MIXED_CONTENT_PASSIVE
  • +12 more
15 rules
GDPR
6
1 err
4 warn
1 notice
  • GDPR_NO_CONSENT_BANNER
  • GDPR_TRACKING_WITHOUT_CONSENT
  • GDPR_THIRD_PARTY_FONTS
  • +3 more
6 rules
GEO Readiness · the wedge
9
0 err
1 warn
8 notice
  • GEO_LOW_ENTITY_DENSITY
  • GEO_NO_ARTICLE_SCHEMA
  • GEO_NO_AUTHOR_ENTITY
  • GEO_NO_FAQ_SCHEMA
  • GEO_NO_LLMS_TXT
  • GEO_SPEAKABLE_MISSING
  • GEO_NO_CITE_ELEMENT
  • GEO_NO_CITE_SIGNALS
  • GEO_NO_EXTERNAL_CITATIONS

The new SEO is the AI search era. Scannica audits how well your pages are read by Google's AI Overviews, ChatGPT, Perplexity — entity density, citation markup, llms.txt, FAQ + Article schema. No incumbent ships these.

9 rules
02 · why local-first

Your crawls aren't anyone else's product.

Cloud auditors built a business on uploading your clients' HTML and renting it back to you by the URL. Scannica isn't an SDK or a thin wrapper around a remote API — the crawler is the binary. Results sit in a local SQLite database. Nothing is phoned home. There's no Scannica server to phone home to.

0 outbound connections to scannica.app during a crawl
Cloud auditor Scannica · desktop
  • Where the data lives
    Their database. They retain the right to look at it.
    A .scannica file on your disk. ZIP of SQLite + config. You can email it.
  • Outbound traffic
    Every URL, every HTML response, every header — uploaded.
    Only the HTTP requests your crawl makes to the site you're auditing.
  • Crawl ceiling
    Per-URL credit, per-month credit, per-seat credit — pick the tax.
    Free 500 URLs · Pro 100K · Enterprise unlimited. Capped on what your hardware can hold.
  • Speed
    Bottlenecked on shared crawl workers and your upload link.
    Rust workers + SQLite WAL. Saturates your bandwidth, not their queue.
  • Pause and resume
    Crawl runs in their pipeline. You watch a progress bar.
    Pause mid-crawl, close the laptop, resume on the train. Session state persists.
  • Offline
    Doesn't work. Even reading old reports needs their server.
    Open last quarter's .scannica file on a plane. Works fine.
  • Telemetry
    App analytics, error reporting, session replay — opt-out at best.
    None. There is no Scannica server to phone home to.
* Honest disclosure: Free tier caps the crawl at 500 URLs. Pro at 100K. Enterprise unlimited. Your hardware sets the practical ceiling on the upper tiers.
03 · a look inside

Dense. Keyboard-first. Built like a tool, not a funnel.

Built for SEOs who already know what they're looking for. Nothing stands between you and the finding. Every screen is filterable, sortable, exportable. One click ships a finding to Jira, GitHub, ClickUp, or Monday.

Captures from build 0.1.x · UI evolves rapidly pre-1.0
Smart Report ·01
01 Smart Report Site-level findings, ranked by impact. Quick Wins are clusters with the highest fix-to-finding ratio. Leads with the delta since last crawl.
Issue aggregation ·02
02 Issue aggregation Every finding rolls up by rule code. Filter by category, severity, or affected URLs. One click pushes a finding to Jira / GitHub / ClickUp.
Per-page audit ·03
03 Per-page audit Numeric scores per category. No letter grades — the dashboard now leads with severity-coloured numbers and deltas vs the previous crawl.
Configure crawl ·04
04 Configure crawl Depth, concurrency, RPS, timeout, user agent, robots.txt respect, regex include / exclude. JS rendering is a Pro tier feature.
· 0 / 0
04 · GEO readiness

Audit for the retrieval-augmented web.

The retrieval-side surface — Google AI Overviews, ChatGPT browsing, Perplexity, Copilot, Claude search — reads HTML differently from a 2015 spider. Entity density, author markup, citation signals, schema breadth, llms.txt. Things no incumbent auditor checks. Scannica does.

9 checks · all category: GEO Readiness
  1. 01

    llms.txt missing

    GEO_NO_LLMS_TXT

    OpenAI, Anthropic, Perplexity look here for crawl preferences. Without it, AI crawlers default to whatever robots.txt says — which usually says yes.

  2. 02

    No author entity markup

    GEO_NO_AUTHOR_ENTITY

    AI Overviews lean on byline + Schema.org Person markup to attribute and trust a passage. A page with no author entity is harder to cite.

  3. 03

    Low entity density

    GEO_LOW_ENTITY_DENSITY

    Pages thick with named entities — products, people, places, dates — are easier for retrieval-augmented systems to index against a query.

  4. 04

    FAQ schema absent

    GEO_NO_FAQ_SCHEMA

    FAQ + Article structured data are the cleanest way for an LLM to extract a Q&A pair verbatim. Without it, the model has to guess.

  5. 05

    Article schema absent

    GEO_NO_ARTICLE_SCHEMA

    Article markup gives the engine date, author, body — the metadata that ends up in the citation card next to your answer.

  6. 06

    No <cite> element

    GEO_NO_CITE_ELEMENT

    Pages that already cite their own sources signal credibility. Engines that cite *you* prefer pages that *cite others*.

  7. 07

    No outbound citations

    GEO_NO_EXTERNAL_CITATIONS

    An island page with zero outbound links to authoritative sources reads, to a model, like an unverifiable claim.

  8. 08

    No citation context

    GEO_NO_CITE_SIGNALS

    Even with outbound links, lack of in-text citation cues ("per", "according to", "source:") makes machine extraction noisier.

  9. 09

    Speakable missing

    GEO_SPEAKABLE_MISSING

    For voice + assistant surfaces, the SpeakableSpecification schema marks the passage that should be read aloud.

05 · built for

Five jobs. One install.

The same crawl + catalog + report shape adapts to whatever role you're running it under. No separate "dev" SKU vs "marketing" SKU.

06 · tiers

Honest caps. No surprise meter.

Pricing isn't finalized — Scannica is pre-1.0. The shape of the tiers, however, is. Free runs every category check on a small site. Pro unlocks GEO, JS rendering, sitemap diff, exports, and integrations. Enterprise removes the caps.

Free

Try the crawl loop end-to-end.

Crawl cap
500 URLs
Audits
10 pages
Saving
SEO + Performance + Accessibility + Best Practices
Security checks (full)
GDPR summary
Carbon totals
Save / open .scannica files
JS rendering
Sitemap diff
GEO Readiness checks
Exports (CSV, JSON, HTML, Sitemap)
Developer integrations
Per-page carbon breakdown

Pro

· primary

The full kit for working SEO and dev teams.

Crawl cap
100K URLs
Audits
1K pages
Saving
.scannica file
Everything in Free, with no category locked
JS rendering (headless browser)
Sitemap diff vs crawl
GEO Readiness — all 9 checks
Full GDPR + fingerprinting risk
Per-page carbon breakdown
Exports: CSV · JSON · HTML report · Sitemap XML
Integrations: Jira · GitHub · Monday · ClickUp · GSC

Enterprise

Volume crawls and license management.

Crawl cap
Unlimited
Audits
Unlimited
Saving
Unlimited
Everything in Pro
Volume crawls (multi-million URL sites)
License management
Priority support (TBA)

* Pricing TBA. Pre-1.0 invite-only — request access via the download section below.

07 · integrations · pro+

Findings, where the work already lives.

The audit isn't done when the report is generated — it's done when somebody fixes the page. One click pushes a finding (with rule code, severity, description, the offending URL) into the tracker the team already opens every morning.

  • J
    Jira
    Create issues from findings
    Pro+
  • G
    GitHub
    Open tickets in any repo
    Pro+
  • C
    ClickUp
    Push to project tasks
    Pro+
  • M
    Monday
    Drop into the right board
    Pro+
  • S
    Search Console
    Pair crawl data with GSC
    Pro+
08 · request access

Pre-1.0. Invite list, not a download button.

Scannica is in a closed pre-release. Builds for macOS, Windows, and Linux exist; we're shipping access in batches as the rule catalog stabilises and the licensing layer hardens. Drop your email and we'll send a build, the install instructions, and a line about what's still in flux.

0 telemetry 0 cloud upload No account required Pause + resume crawls