Help & Documentation

Comprehensive guide to understanding website metrics, technical indicators, and optimization recommendations. Master the tools to improve your search visibility.

Quick Start Guide

This reference covers all metrics used in the Analyzer and Site Audit tools. Each section includes detailed explanations, recommended values, and actionable optimization tips.

Prefer structured learning? Check out the complete Webmaster Guide with step-by-step tutorials and best practices.

Analyzer Metrics Reference

Detailed explanation of every metric displayed in the page analyzer tool.

Overall Score

The overall score is a weighted calculation (0-100) that evaluates multiple aspects of your page:

  • Title & Meta: Presence and quality of title tags and meta descriptions
  • Headings: Proper heading structure (H1-H6) hierarchy
  • Content Quality: Word count, readability, and text/HTML ratio
  • Technical SEO: HTTPS, canonical URLs, robots directives, security headers
  • Social Media: Open Graph and Twitter Card implementation
  • Internal Links: Presence of contextual internal linking
  • Performance: Page load speed and resource optimization
Note: This score is a directional indicator for optimization opportunities—not a search ranking prediction.

Page Information

  • URL: The analyzed URL (after redirects if any)
  • Status Code: HTTP response code (200 = OK, 301 = Permanent Redirect, 404 = Not Found, 500 = Server Error)
  • Final URL: The destination after following redirects
  • Content Type: MIME type of the response (typically text/html for web pages)
  • Render Time: Time taken to fetch and optionally render the page
Best Practice: Ensure pages return 200 status codes. Minimize redirect chains (max 1-2 redirects).

Title & Meta Information

Title Tag
  • Recommended Length: 50-60 characters (~600px)
  • Purpose: Appears in search results, browser tabs, and social shares
  • Best Practices:
    • Include primary keyword near the beginning
    • Make it unique for every page
    • Write for humans, not just search engines
    • Match search intent (informational, transactional, navigational)
    • Include brand name (usually at the end)
Meta Description
  • Recommended Length: 120-160 characters (~960px)
  • Purpose: Summarizes page content in search results; influences click-through rate (CTR)
  • Best Practices:
    • Write compelling, action-oriented copy
    • Include a call-to-action when appropriate
    • Match the content on the page
    • Make it unique for every page
    • Not a direct ranking factor, but affects CTR
Canonical URL
  • Purpose: Specifies the preferred version of a page when duplicates exist
  • Usage:
    • Set self-referential canonical on original pages
    • Point duplicate/variant pages to the canonical version
    • Use absolute URLs (https://example.com/page)
    • Ensure canonical URLs are indexable (not noindex, 404, or redirecting)
Tip: Avoid duplicating titles/descriptions across pages. Each page should have unique metadata aligned with its specific content and target keywords.

Heading Structure

  • H1: Use exactly one H1 per page as the main heading; should be descriptive and include primary keyword
  • H2-H6: Create logical hierarchy; H2 for main sections, H3 for subsections, etc.
  • Best Practices:
    • Don't skip levels (e.g., don't jump from H2 to H4)
    • Make headings descriptive and scannable
    • Use headings to structure content, not for styling
    • Include keywords naturally in headings
    • Headings help users and search engines understand page structure

Content Statistics

  • Word Count: Total visible text words on the page
    • Minimum: 250-300 words for most pages
    • Thin content (<250 words) may underperform in search
    • Focus on satisfying user intent, not arbitrary word counts
    • Longer content (1,000-2,000+ words) often ranks better for competitive topics
  • Character Count: Total characters in visible text
  • Sentence Count: Number of sentences detected
  • Paragraph Count: Number of paragraph blocks
  • Reading Time: Estimated time to read based on ~200 words/minute
  • Text/HTML Ratio: Percentage of text vs. total HTML
    • Low ratios (<15%) may indicate heavy templates or thin content
    • Target: 20-70% depending on page type
    • Add meaningful, user-focused copy to improve ratio

Readability Metrics

  • Flesch Reading Ease: Scale of 0-100 (higher = easier)
    • 90-100: Very Easy (5th grade)
    • 80-89: Easy (6th grade)
    • 70-79: Fairly Easy (7th grade)
    • 60-69: Standard (8th-9th grade) — Target for most content
    • 50-59: Fairly Difficult (10th-12th grade)
    • 30-49: Difficult (College level)
    • 0-29: Very Difficult (Professional/Academic)
  • Flesch-Kincaid Grade Level: U.S. school grade level needed to understand the text
    • Target: 8-10 for general audiences
    • Adjust based on your audience and topic complexity
    • Technical/academic content may naturally score higher
  • SMOG Index: Simple Measure of Gobbledygook; estimates years of education needed
  • Coleman-Liau Index: Grade level based on characters per word and words per sentence
Tip: Aim for clarity without oversimplifying. Use shorter sentences, active voice, and familiar words. Break up text with headings, lists, and images.

SEO Recommendations

Automated suggestions based on detected issues and optimization opportunities:

  • Missing or short title/meta tags
  • Heading structure problems
  • Thin or duplicate content
  • Missing security headers
  • Non-HTTPS resources (mixed content)
  • Missing canonical tags
  • Incorrect robots directives
  • Performance bottlenecks

Prioritize fixes by impact: Critical issues (indexability, status codes) first, then content/meta, then performance.

Technical SEO

Robots & Indexability
  • Robots Meta Tag: Controls indexing (index/noindex) and link following (follow/nofollow)
    • Default: index, follow
    • Use noindex for pages you don't want in search (thank-you pages, admin areas, duplicates)
    • Use nofollow to prevent passing link equity
  • X-Robots-Tag: HTTP header alternative to robots meta tag
  • robots.txt: Controls crawling (not indexing)
    • Don't block pages you want indexed
    • Use to manage crawl budget and protect sensitive areas
    • Test with Google Search Console
  • Indexable: True when status=200 and no noindex directive present
Security Headers
  • HSTS (Strict-Transport-Security): Forces HTTPS; add max-age=31536000; includeSubDomains
  • CSP (Content-Security-Policy): Mitigates XSS attacks; start with report-only mode
  • X-Content-Type-Options: Set to "nosniff" to prevent MIME sniffing
  • X-Frame-Options: Prevents clickjacking; set to DENY or SAMEORIGIN
  • Referrer-Policy: Controls referrer information; consider "strict-origin-when-cross-origin"
  • Permissions-Policy: Restricts browser features (camera, microphone, geolocation)
HTTPS & Mixed Content
  • HTTPS: Essential for trust, security, and ranking
  • Mixed Content: Loading HTTP resources on HTTPS pages triggers browser warnings
  • Fix: Replace all http:// URLs with https:// or use protocol-relative URLs (//)
Common Pitfall: Using robots.txt to block pages you want indexed. Use noindex meta tag instead, and allow crawling.

Social Meta Tags

Open Graph (Facebook, LinkedIn, etc.)
  • og:title: Title for social shares (can differ from page title)
  • og:description: Description for social previews
  • og:image: Image URL for social cards (1200x630px recommended)
  • og:url: Canonical URL for the page
  • og:type: Type of content (website, article, product, etc.)
  • og:site_name: Name of your website
Twitter Cards
  • twitter:card: Card type (summary, summary_large_image, app, player)
  • twitter:site: @username of website
  • twitter:creator: @username of content creator
  • twitter:title: Title for Twitter
  • twitter:description: Description for Twitter
  • twitter:image: Image for Twitter card

Hreflang (Internationalization)

  • Purpose: Tells search engines which language/region versions exist
  • Format: Use BCP-47 language codes (en, en-GB, en-US, es, fr, de, etc.)
  • Implementation:
    • HTML link tags in <head>
    • HTTP headers (for non-HTML content)
    • XML sitemap
  • Best Practices:
    • Use reciprocal links (every alternate page must link back)
    • Include x-default for fallback/language selector
    • Use absolute URLs
    • Link to the canonical version of each language
    • Test in Google Search Console
  • Internal Links: Links to pages on the same domain
    • Build topic clusters and content hubs
    • Use descriptive anchor text
    • Link to cornerstone content
    • Ensure important pages are 1-3 clicks from homepage
  • External Links: Links to other domains
    • Cite authoritative sources
    • Keep relevant to your content
    • Add rel="noopener noreferrer" when using target="_blank"
  • rel Attributes:
    • nofollow: Don't pass link equity (use for untrusted content, paid links)
    • ugc: User-generated content links
    • sponsored: Paid/advertising links
    • noopener: Security for target="_blank" links
    • noreferrer: Don't send referrer information
  • Broken Links: Links returning 4xx/5xx errors; fix or remove
  • Redirects: Minimize redirect chains; update links to point to final destination

Keyword Density

Shows the most frequent terms and phrases on your page:

  • 1-word: Individual terms and their frequency
  • 2-word: Bigrams (two-word phrases)
  • 3-word: Trigrams (three-word phrases)

Best Practices:

  • Use keywords naturally; avoid stuffing (excessive repetition)
  • Target density: 1-3% for primary keywords (not a strict rule)
  • Focus on topical relevance and semantic variety
  • Include synonyms and related terms (LSI keywords)
  • Align page terms with target search queries
Warning: Keyword density is not a ranking factor. Write for humans first, optimize for keywords naturally.

Structured Data (Schema.org)

Structured data helps search engines understand your content and enables rich results:

  • Format: JSON-LD (recommended), Microdata, or RDFa
  • Common Types:
    • Organization: Company info, logo, social profiles
    • Article: Blog posts, news articles
    • Product: E-commerce products with price, availability, reviews
    • Review: Ratings and reviews
    • Event: Concerts, webinars, conferences
    • FAQ: Frequently asked questions
    • HowTo: Step-by-step instructions
    • Breadcrumb: Navigation path
    • LocalBusiness: Physical locations
    • Person: Author profiles
  • Benefits:
    • Rich results in search (star ratings, images, prices)
    • Knowledge Graph eligibility
    • Voice search optimization
    • Enhanced visibility and CTR
  • Validation: Use our Schema Checker or schema.org validator
Powered by Extruct: Our analyzer uses the Extruct library to extract all structured data formats from your pages.

Keyword Suggestions (AI-Powered)

AI-powered keyword extraction using KeyBERT to find relevant topics and phrases:

  • Technology: KeyBERT with sentence-transformers (BERT embeddings)
  • Metrics:
    • Rank: Relevance ranking (1 = most relevant)
    • Relevance Score: Similarity to document content (0-1)
    • Occurrences: Number of times phrase appears
    • Coverage: Percentage of document where phrase appears
  • Use Cases:
    • Discover main topics and themes
    • Find content gaps and expansion opportunities
    • Validate keyword targeting
    • Identify semantic relevance
    • Optimize existing content
Powered by KeyBERT: State-of-the-art keyword extraction using BERT contextual embeddings for semantic understanding.

Advanced Diagnostics

Images Analysis
  • Total Images: Count of <img> tags on page
  • Missing Alt Text: Images without alt attributes (accessibility + SEO issue)
  • Broken Images: Images that failed to load (4xx/5xx errors)
  • Best Practices:
    • Add descriptive alt text to all images
    • Use alt="" for decorative images
    • Optimize image file sizes (compression, modern formats like WebP)
    • Use responsive images (srcset, sizes attributes)
    • Lazy load off-screen images
Links Analysis
  • Total Links: All <a> tags on the page
  • Internal vs. External: Distribution of link destinations
  • Follow vs. Nofollow: Link equity flow analysis
  • Broken Links: Links to 404 pages or dead endpoints
  • Unsafe Targets: target="_blank" without rel="noopener noreferrer"
Canonical Validation
  • Checks if canonical URL is properly set
  • Verifies canonical points to indexable page
  • Detects canonical chain issues
  • Flags cross-domain canonicals

PageSpeed Insights

Real-time performance analysis powered by Google PageSpeed Insights API:

Core Web Vitals
  • LCP (Largest Contentful Paint): Loading performance
    • Good: < 2.5s
    • Needs Improvement: 2.5-4.0s
    • Poor: > 4.0s
  • FID (First Input Delay): Interactivity
    • Good: < 100ms
    • Needs Improvement: 100-300ms
    • Poor: > 300ms
  • CLS (Cumulative Layout Shift): Visual stability
    • Good: < 0.1
    • Needs Improvement: 0.1-0.25
    • Poor: > 0.25
Other Metrics
  • FCP (First Contentful Paint): Time to first visible content
  • SI (Speed Index): How quickly content is visually displayed
  • TTI (Time to Interactive): When page becomes fully interactive
  • TBT (Total Blocking Time): Time main thread is blocked
  • TTFB (Time to First Byte): Server response time
Optimization Opportunities
  • Reduce unused JavaScript and CSS
  • Optimize images (compression, modern formats, lazy loading)
  • Minify and compress resources
  • Enable browser caching
  • Use a Content Delivery Network (CDN)
  • Eliminate render-blocking resources
  • Reduce server response time
  • Avoid large layout shifts
Note: Core Web Vitals are confirmed ranking factors. Improving these metrics can enhance user experience and search visibility.

Site Audit Reference

Understanding the site audit tool and its findings.

Audit Overview

The Site Audit tool crawls your website (up to a specified depth and page limit) to identify technical and content issues:

  • Crawl Depth: Maximum number of clicks from the starting URL
  • Max Pages: Limit on total pages to crawl
  • Render JS: Option to execute JavaScript (slower but necessary for SPAs)
  • Respect Noindex: Option to skip pages marked noindex

Audit Summary

Aggregated counts of issues across all crawled pages:

  • Total pages crawled
  • HTTP errors (4xx, 5xx)
  • Missing titles and H1s
  • Thin content pages
  • Duplicate title/content clusters
  • Security header issues
  • Noindexed pages

Audit Issue Codes

  • TITLE_MISSING: Page lacks a <title> tag
  • TITLE_SHORT: Title is too short (<30 characters)
  • TITLE_LONG: Title exceeds 60 characters
  • H1_MISSING: No H1 heading found
  • H1_MULTIPLE: More than one H1 (may be intentional, but review structure)
  • THIN_CONTENT: Word count below 250 words
  • HSTS_MISSING: HTTPS site without HSTS header
  • NOINDEX: Page marked noindex (verify intentional)
  • HTTP_ERROR: Non-200 status code (404, 500, etc.)
  • REDIRECT: Page returns 301/302 redirect
  • DUPLICATE_TITLE: Title is very similar to another page
  • DUPLICATE_CONTENT: Body content matches another page closely
  • MIXED_CONTENT: Loading HTTP resources on HTTPS page

Pages Table

Detailed view of each crawled page:

  • URL: Full page address
  • Status: HTTP response code
  • Title: Page title tag
  • Word Count: Visible text word count
  • H1 Count: Number of H1 headings
  • Internal Links: Count of internal links
  • External Links: Count of external links
  • Noindex: Whether page requests exclusion from indexing
  • Issues: List of detected issue codes

Visualizes internal linking structure:

  • Nodes: Each crawled page
  • Edges: Internal links between pages
  • Inbound Links: Pages linking to this URL
  • Outbound Links: Pages this URL links to
  • Click Depth: Distance from starting URL
  • Orphan Pages: Pages with zero internal inbound links

Best Practices:

  • Keep important pages within 3 clicks of homepage
  • Eliminate orphan pages (add internal links)
  • Create hub pages with strong internal linking
  • Build topic clusters with pillar content

Duplicate Clusters

Groups pages with similar titles or content:

  • Title Duplicates: RapidFuzz token-set similarity ≥ 92%
  • Content Duplicates: Simhash Hamming distance ≤ 3 bits

Actions to Take:

  • Canonicalize: Point duplicates to the preferred version
  • Consolidate: Merge thin duplicate pages into comprehensive resource
  • Differentiate: Rewrite to make each page unique and valuable
  • Noindex: Exclude low-value duplicates from indexing
  • 301 Redirect: Redirect obsolete pages to active versions

Sitemap Generator

Creates XML sitemaps for search engine submission:

  • Respect Noindex: Exclude pages marked noindex
  • Use Canonical URLs: Include only canonical versions
  • Coverage Analysis: Compare sitemap URLs vs. crawled pages
  • Lastmod: Include last modification date (if available)
  • Priority: Assign priority values (0.0-1.0)
  • Changefreq: Indicate update frequency

Best Practices:

  • Submit sitemap to Google Search Console and Bing Webmaster Tools
  • Keep sitemaps under 50MB and 50,000 URLs (split if needed)
  • Update sitemaps when content changes
  • Include only indexable, canonical URLs
  • Use sitemap index for multiple sitemaps

Advanced Features

JavaScript Rendering

Enable "Render JS" to analyze Single Page Applications (SPAs) and JavaScript-heavy sites:

  • When to Use:
    • React, Vue, Angular, Svelte applications
    • Client-side rendered content
    • Dynamic content loaded via AJAX
    • Content behind JavaScript interactions
  • Technology: Playwright headless browser automation
  • Trade-offs: Slower analysis but sees content as users and search engines do

Share Reports

Generate shareable links to analysis results:

  • Set expiration time (7 days, 30 days, etc.)
  • Share with clients, team members, or stakeholders
  • No login required to view shared reports

Local Report Storage

Save analysis reports in your browser:

  • Stored locally using browser Local Storage
  • No data sent to servers (privacy-friendly)
  • Access saved reports from "Reports List"
  • Compare historical analysis

Frequently Asked Questions

Why isn't my score 100?

The overall score balances multiple factors: title/meta quality, heading structure, content depth, internal links, technical SEO, security headers, and performance. It's a directional signal for optimization opportunities—not a search ranking prediction. Even well-optimized pages may score 70-90 due to subjective factors.

Do meta keywords still matter?

No. Meta keywords tags are ignored by all major search engines (Google, Bing, Yahoo). Focus on title tags, headings, and high-quality content instead.

When should I use "Render JS"?

Enable JavaScript rendering when analyzing SPAs (React, Vue, Angular) or pages where critical content loads via JavaScript. It's slower but ensures you see what search engines and users see after client-side rendering.

How accurate is the PageSpeed analysis?

PageSpeed results come directly from Google's PageSpeed Insights API, which uses real Chrome User Experience Report (CrUX) data when available and Lighthouse lab testing. Results may vary based on location, device, network, and timing.

Can I analyze password-protected pages?

Not currently. The analyzer can only access publicly available pages. For staging sites, consider using HTTP authentication that can be passed in the URL or temporarily allowing access by IP.

Why are some recommendations conflicting?

Some optimizations involve trade-offs. For example, longer content improves depth but may hurt readability scores. Use professional judgment to balance recommendations based on your goals and audience.

How often should I audit my site?

Run audits:

  • Monthly for active sites
  • After major updates or redesigns
  • When launching new content sections
  • Before and after technical changes
  • When investigating ranking drops
What's the difference between noindex and robots.txt blocking?

robots.txt controls crawling (whether bots can access the page). noindex controls indexing (whether the page can appear in search results). To keep a page out of search results, use noindex and allow crawling. Blocking with robots.txt may prevent search engines from seeing the noindex directive.

Why does my page have duplicate content issues?

Common causes:

  • Multiple URLs for the same content (with/without www, http/https, trailing slashes)
  • Printer-friendly versions
  • URL parameters creating variations
  • Syndicated or scraped content
  • Product pages with similar descriptions

Solutions: Implement canonical tags, use 301 redirects, consolidate similar pages, or use noindex for duplicates.

Quick Optimization Checklist
  • Unique, descriptive title (50-60 characters) with primary keyword
  • Compelling meta description (120-160 characters)
  • Clear heading structure (one H1, logical H2-H6 hierarchy)
  • Sufficient content depth (minimum 300 words for most pages)
  • Good readability (Flesch score 60-70 for general audiences)
  • Descriptive internal links to key pages
  • Canonical tags set correctly
  • Proper robots directives (noindex only where intended)
  • HTTPS enabled with HSTS header
  • Security headers (CSP, X-Content-Type-Options, X-Frame-Options)
  • Open Graph and Twitter Card tags
  • Structured data (JSON-LD) for relevant content types
  • Alt text on all images
  • No broken links (404s)
  • Core Web Vitals in "Good" range (LCP < 2.5s, FID < 100ms, CLS < 0.1)
  • Mobile-friendly and responsive design
  • XML sitemap submitted to search engines
  • Important pages within 3 clicks of homepage

Ready to optimize your website?

Analyze Your Site Now