Help & Documentation
Comprehensive guide to understanding website metrics, technical indicators, and optimization recommendations. Master the tools to improve your search visibility.
Quick Start Guide
This reference covers all metrics used in the Analyzer and Site Audit tools. Each section includes detailed explanations, recommended values, and actionable optimization tips.
Prefer structured learning? Check out the complete Webmaster Guide with step-by-step tutorials and best practices.
Analyzer Metrics Reference
Detailed explanation of every metric displayed in the page analyzer tool.
Overall Score
The overall score is a weighted calculation (0-100) that evaluates multiple aspects of your page:
- Title & Meta: Presence and quality of title tags and meta descriptions
- Headings: Proper heading structure (H1-H6) hierarchy
- Content Quality: Word count, readability, and text/HTML ratio
- Technical SEO: HTTPS, canonical URLs, robots directives, security headers
- Social Media: Open Graph and Twitter Card implementation
- Internal Links: Presence of contextual internal linking
- Performance: Page load speed and resource optimization
Page Information
- URL: The analyzed URL (after redirects if any)
- Status Code: HTTP response code (200 = OK, 301 = Permanent Redirect, 404 = Not Found, 500 = Server Error)
- Final URL: The destination after following redirects
- Content Type: MIME type of the response (typically text/html for web pages)
- Render Time: Time taken to fetch and optionally render the page
Title & Meta Information
Title Tag
- Recommended Length: 50-60 characters (~600px)
- Purpose: Appears in search results, browser tabs, and social shares
- Best Practices:
- Include primary keyword near the beginning
- Make it unique for every page
- Write for humans, not just search engines
- Match search intent (informational, transactional, navigational)
- Include brand name (usually at the end)
Meta Description
- Recommended Length: 120-160 characters (~960px)
- Purpose: Summarizes page content in search results; influences click-through rate (CTR)
- Best Practices:
- Write compelling, action-oriented copy
- Include a call-to-action when appropriate
- Match the content on the page
- Make it unique for every page
- Not a direct ranking factor, but affects CTR
Canonical URL
- Purpose: Specifies the preferred version of a page when duplicates exist
- Usage:
- Set self-referential canonical on original pages
- Point duplicate/variant pages to the canonical version
- Use absolute URLs (https://example.com/page)
- Ensure canonical URLs are indexable (not noindex, 404, or redirecting)
Heading Structure
- H1: Use exactly one H1 per page as the main heading; should be descriptive and include primary keyword
- H2-H6: Create logical hierarchy; H2 for main sections, H3 for subsections, etc.
- Best Practices:
- Don't skip levels (e.g., don't jump from H2 to H4)
- Make headings descriptive and scannable
- Use headings to structure content, not for styling
- Include keywords naturally in headings
- Headings help users and search engines understand page structure
Content Statistics
- Word Count: Total visible text words on the page
- Minimum: 250-300 words for most pages
- Thin content (<250 words) may underperform in search
- Focus on satisfying user intent, not arbitrary word counts
- Longer content (1,000-2,000+ words) often ranks better for competitive topics
- Character Count: Total characters in visible text
- Sentence Count: Number of sentences detected
- Paragraph Count: Number of paragraph blocks
- Reading Time: Estimated time to read based on ~200 words/minute
- Text/HTML Ratio: Percentage of text vs. total HTML
- Low ratios (<15%) may indicate heavy templates or thin content
- Target: 20-70% depending on page type
- Add meaningful, user-focused copy to improve ratio
Readability Metrics
- Flesch Reading Ease: Scale of 0-100 (higher = easier)
- 90-100: Very Easy (5th grade)
- 80-89: Easy (6th grade)
- 70-79: Fairly Easy (7th grade)
- 60-69: Standard (8th-9th grade) — Target for most content
- 50-59: Fairly Difficult (10th-12th grade)
- 30-49: Difficult (College level)
- 0-29: Very Difficult (Professional/Academic)
- Flesch-Kincaid Grade Level: U.S. school grade level needed to understand the text
- Target: 8-10 for general audiences
- Adjust based on your audience and topic complexity
- Technical/academic content may naturally score higher
- SMOG Index: Simple Measure of Gobbledygook; estimates years of education needed
- Coleman-Liau Index: Grade level based on characters per word and words per sentence
SEO Recommendations
Automated suggestions based on detected issues and optimization opportunities:
- Missing or short title/meta tags
- Heading structure problems
- Thin or duplicate content
- Missing security headers
- Non-HTTPS resources (mixed content)
- Missing canonical tags
- Incorrect robots directives
- Performance bottlenecks
Prioritize fixes by impact: Critical issues (indexability, status codes) first, then content/meta, then performance.
Technical SEO
Robots & Indexability
- Robots Meta Tag: Controls indexing (index/noindex) and link following (follow/nofollow)
- Default: index, follow
- Use noindex for pages you don't want in search (thank-you pages, admin areas, duplicates)
- Use nofollow to prevent passing link equity
- X-Robots-Tag: HTTP header alternative to robots meta tag
- robots.txt: Controls crawling (not indexing)
- Don't block pages you want indexed
- Use to manage crawl budget and protect sensitive areas
- Test with Google Search Console
- Indexable: True when status=200 and no noindex directive present
Security Headers
- HSTS (Strict-Transport-Security): Forces HTTPS; add max-age=31536000; includeSubDomains
- CSP (Content-Security-Policy): Mitigates XSS attacks; start with report-only mode
- X-Content-Type-Options: Set to "nosniff" to prevent MIME sniffing
- X-Frame-Options: Prevents clickjacking; set to DENY or SAMEORIGIN
- Referrer-Policy: Controls referrer information; consider "strict-origin-when-cross-origin"
- Permissions-Policy: Restricts browser features (camera, microphone, geolocation)
HTTPS & Mixed Content
- HTTPS: Essential for trust, security, and ranking
- Mixed Content: Loading HTTP resources on HTTPS pages triggers browser warnings
- Fix: Replace all http:// URLs with https:// or use protocol-relative URLs (//)
Social Meta Tags
Open Graph (Facebook, LinkedIn, etc.)
- og:title: Title for social shares (can differ from page title)
- og:description: Description for social previews
- og:image: Image URL for social cards (1200x630px recommended)
- og:url: Canonical URL for the page
- og:type: Type of content (website, article, product, etc.)
- og:site_name: Name of your website
Twitter Cards
- twitter:card: Card type (summary, summary_large_image, app, player)
- twitter:site: @username of website
- twitter:creator: @username of content creator
- twitter:title: Title for Twitter
- twitter:description: Description for Twitter
- twitter:image: Image for Twitter card
Hreflang (Internationalization)
- Purpose: Tells search engines which language/region versions exist
- Format: Use BCP-47 language codes (en, en-GB, en-US, es, fr, de, etc.)
- Implementation:
- HTML link tags in <head>
- HTTP headers (for non-HTML content)
- XML sitemap
- Best Practices:
- Use reciprocal links (every alternate page must link back)
- Include x-default for fallback/language selector
- Use absolute URLs
- Link to the canonical version of each language
- Test in Google Search Console
Link Attributes & Analysis
- Internal Links: Links to pages on the same domain
- Build topic clusters and content hubs
- Use descriptive anchor text
- Link to cornerstone content
- Ensure important pages are 1-3 clicks from homepage
- External Links: Links to other domains
- Cite authoritative sources
- Keep relevant to your content
- Add rel="noopener noreferrer" when using target="_blank"
- rel Attributes:
- nofollow: Don't pass link equity (use for untrusted content, paid links)
- ugc: User-generated content links
- sponsored: Paid/advertising links
- noopener: Security for target="_blank" links
- noreferrer: Don't send referrer information
- Broken Links: Links returning 4xx/5xx errors; fix or remove
- Redirects: Minimize redirect chains; update links to point to final destination
Keyword Density
Shows the most frequent terms and phrases on your page:
- 1-word: Individual terms and their frequency
- 2-word: Bigrams (two-word phrases)
- 3-word: Trigrams (three-word phrases)
Best Practices:
- Use keywords naturally; avoid stuffing (excessive repetition)
- Target density: 1-3% for primary keywords (not a strict rule)
- Focus on topical relevance and semantic variety
- Include synonyms and related terms (LSI keywords)
- Align page terms with target search queries
Structured Data (Schema.org)
Structured data helps search engines understand your content and enables rich results:
- Format: JSON-LD (recommended), Microdata, or RDFa
- Common Types:
- Organization: Company info, logo, social profiles
- Article: Blog posts, news articles
- Product: E-commerce products with price, availability, reviews
- Review: Ratings and reviews
- Event: Concerts, webinars, conferences
- FAQ: Frequently asked questions
- HowTo: Step-by-step instructions
- Breadcrumb: Navigation path
- LocalBusiness: Physical locations
- Person: Author profiles
- Benefits:
- Rich results in search (star ratings, images, prices)
- Knowledge Graph eligibility
- Voice search optimization
- Enhanced visibility and CTR
- Validation: Use our Schema Checker or schema.org validator
Keyword Suggestions (AI-Powered)
AI-powered keyword extraction using KeyBERT to find relevant topics and phrases:
- Technology: KeyBERT with sentence-transformers (BERT embeddings)
- Metrics:
- Rank: Relevance ranking (1 = most relevant)
- Relevance Score: Similarity to document content (0-1)
- Occurrences: Number of times phrase appears
- Coverage: Percentage of document where phrase appears
- Use Cases:
- Discover main topics and themes
- Find content gaps and expansion opportunities
- Validate keyword targeting
- Identify semantic relevance
- Optimize existing content
Advanced Diagnostics
Images Analysis
- Total Images: Count of <img> tags on page
- Missing Alt Text: Images without alt attributes (accessibility + SEO issue)
- Broken Images: Images that failed to load (4xx/5xx errors)
- Best Practices:
- Add descriptive alt text to all images
- Use alt="" for decorative images
- Optimize image file sizes (compression, modern formats like WebP)
- Use responsive images (srcset, sizes attributes)
- Lazy load off-screen images
Links Analysis
- Total Links: All <a> tags on the page
- Internal vs. External: Distribution of link destinations
- Follow vs. Nofollow: Link equity flow analysis
- Broken Links: Links to 404 pages or dead endpoints
- Unsafe Targets: target="_blank" without rel="noopener noreferrer"
Canonical Validation
- Checks if canonical URL is properly set
- Verifies canonical points to indexable page
- Detects canonical chain issues
- Flags cross-domain canonicals
PageSpeed Insights
Real-time performance analysis powered by Google PageSpeed Insights API:
Core Web Vitals
- LCP (Largest Contentful Paint): Loading performance
- Good: < 2.5s
- Needs Improvement: 2.5-4.0s
- Poor: > 4.0s
- FID (First Input Delay): Interactivity
- Good: < 100ms
- Needs Improvement: 100-300ms
- Poor: > 300ms
- CLS (Cumulative Layout Shift): Visual stability
- Good: < 0.1
- Needs Improvement: 0.1-0.25
- Poor: > 0.25
Other Metrics
- FCP (First Contentful Paint): Time to first visible content
- SI (Speed Index): How quickly content is visually displayed
- TTI (Time to Interactive): When page becomes fully interactive
- TBT (Total Blocking Time): Time main thread is blocked
- TTFB (Time to First Byte): Server response time
Optimization Opportunities
- Reduce unused JavaScript and CSS
- Optimize images (compression, modern formats, lazy loading)
- Minify and compress resources
- Enable browser caching
- Use a Content Delivery Network (CDN)
- Eliminate render-blocking resources
- Reduce server response time
- Avoid large layout shifts
Site Audit Reference
Understanding the site audit tool and its findings.
Audit Overview
The Site Audit tool crawls your website (up to a specified depth and page limit) to identify technical and content issues:
- Crawl Depth: Maximum number of clicks from the starting URL
- Max Pages: Limit on total pages to crawl
- Render JS: Option to execute JavaScript (slower but necessary for SPAs)
- Respect Noindex: Option to skip pages marked noindex
Audit Summary
Aggregated counts of issues across all crawled pages:
- Total pages crawled
- HTTP errors (4xx, 5xx)
- Missing titles and H1s
- Thin content pages
- Duplicate title/content clusters
- Security header issues
- Noindexed pages
Audit Issue Codes
- TITLE_MISSING: Page lacks a <title> tag
- TITLE_SHORT: Title is too short (<30 characters)
- TITLE_LONG: Title exceeds 60 characters
- H1_MISSING: No H1 heading found
- H1_MULTIPLE: More than one H1 (may be intentional, but review structure)
- THIN_CONTENT: Word count below 250 words
- HSTS_MISSING: HTTPS site without HSTS header
- NOINDEX: Page marked noindex (verify intentional)
- HTTP_ERROR: Non-200 status code (404, 500, etc.)
- REDIRECT: Page returns 301/302 redirect
- DUPLICATE_TITLE: Title is very similar to another page
- DUPLICATE_CONTENT: Body content matches another page closely
- MIXED_CONTENT: Loading HTTP resources on HTTPS page
Pages Table
Detailed view of each crawled page:
- URL: Full page address
- Status: HTTP response code
- Title: Page title tag
- Word Count: Visible text word count
- H1 Count: Number of H1 headings
- Internal Links: Count of internal links
- External Links: Count of external links
- Noindex: Whether page requests exclusion from indexing
- Issues: List of detected issue codes
Link Graph
Visualizes internal linking structure:
- Nodes: Each crawled page
- Edges: Internal links between pages
- Inbound Links: Pages linking to this URL
- Outbound Links: Pages this URL links to
- Click Depth: Distance from starting URL
- Orphan Pages: Pages with zero internal inbound links
Best Practices:
- Keep important pages within 3 clicks of homepage
- Eliminate orphan pages (add internal links)
- Create hub pages with strong internal linking
- Build topic clusters with pillar content
Duplicate Clusters
Groups pages with similar titles or content:
- Title Duplicates: RapidFuzz token-set similarity ≥ 92%
- Content Duplicates: Simhash Hamming distance ≤ 3 bits
Actions to Take:
- Canonicalize: Point duplicates to the preferred version
- Consolidate: Merge thin duplicate pages into comprehensive resource
- Differentiate: Rewrite to make each page unique and valuable
- Noindex: Exclude low-value duplicates from indexing
- 301 Redirect: Redirect obsolete pages to active versions
Sitemap Generator
Creates XML sitemaps for search engine submission:
- Respect Noindex: Exclude pages marked noindex
- Use Canonical URLs: Include only canonical versions
- Coverage Analysis: Compare sitemap URLs vs. crawled pages
- Lastmod: Include last modification date (if available)
- Priority: Assign priority values (0.0-1.0)
- Changefreq: Indicate update frequency
Best Practices:
- Submit sitemap to Google Search Console and Bing Webmaster Tools
- Keep sitemaps under 50MB and 50,000 URLs (split if needed)
- Update sitemaps when content changes
- Include only indexable, canonical URLs
- Use sitemap index for multiple sitemaps
Advanced Features
JavaScript Rendering
Enable "Render JS" to analyze Single Page Applications (SPAs) and JavaScript-heavy sites:
- When to Use:
- React, Vue, Angular, Svelte applications
- Client-side rendered content
- Dynamic content loaded via AJAX
- Content behind JavaScript interactions
- Technology: Playwright headless browser automation
- Trade-offs: Slower analysis but sees content as users and search engines do
Share Reports
Generate shareable links to analysis results:
- Set expiration time (7 days, 30 days, etc.)
- Share with clients, team members, or stakeholders
- No login required to view shared reports
Local Report Storage
Save analysis reports in your browser:
- Stored locally using browser Local Storage
- No data sent to servers (privacy-friendly)
- Access saved reports from "Reports List"
- Compare historical analysis
Frequently Asked Questions
Why isn't my score 100?
The overall score balances multiple factors: title/meta quality, heading structure, content depth, internal links, technical SEO, security headers, and performance. It's a directional signal for optimization opportunities—not a search ranking prediction. Even well-optimized pages may score 70-90 due to subjective factors.
Do meta keywords still matter?
No. Meta keywords tags are ignored by all major search engines (Google, Bing, Yahoo). Focus on title tags, headings, and high-quality content instead.
When should I use "Render JS"?
Enable JavaScript rendering when analyzing SPAs (React, Vue, Angular) or pages where critical content loads via JavaScript. It's slower but ensures you see what search engines and users see after client-side rendering.
How accurate is the PageSpeed analysis?
PageSpeed results come directly from Google's PageSpeed Insights API, which uses real Chrome User Experience Report (CrUX) data when available and Lighthouse lab testing. Results may vary based on location, device, network, and timing.
Can I analyze password-protected pages?
Not currently. The analyzer can only access publicly available pages. For staging sites, consider using HTTP authentication that can be passed in the URL or temporarily allowing access by IP.
Why are some recommendations conflicting?
Some optimizations involve trade-offs. For example, longer content improves depth but may hurt readability scores. Use professional judgment to balance recommendations based on your goals and audience.
How often should I audit my site?
Run audits:
- Monthly for active sites
- After major updates or redesigns
- When launching new content sections
- Before and after technical changes
- When investigating ranking drops
What's the difference between noindex and robots.txt blocking?
robots.txt controls crawling (whether bots can access the page). noindex controls indexing (whether the page can appear in search results). To keep a page out of search results, use noindex and allow crawling. Blocking with robots.txt may prevent search engines from seeing the noindex directive.
Why does my page have duplicate content issues?
Common causes:
- Multiple URLs for the same content (with/without www, http/https, trailing slashes)
- Printer-friendly versions
- URL parameters creating variations
- Syndicated or scraped content
- Product pages with similar descriptions
Solutions: Implement canonical tags, use 301 redirects, consolidate similar pages, or use noindex for duplicates.
Quick Optimization Checklist
- ✓ Unique, descriptive title (50-60 characters) with primary keyword
- ✓ Compelling meta description (120-160 characters)
- ✓ Clear heading structure (one H1, logical H2-H6 hierarchy)
- ✓ Sufficient content depth (minimum 300 words for most pages)
- ✓ Good readability (Flesch score 60-70 for general audiences)
- ✓ Descriptive internal links to key pages
- ✓ Canonical tags set correctly
- ✓ Proper robots directives (noindex only where intended)
- ✓ HTTPS enabled with HSTS header
- ✓ Security headers (CSP, X-Content-Type-Options, X-Frame-Options)
- ✓ Open Graph and Twitter Card tags
- ✓ Structured data (JSON-LD) for relevant content types
- ✓ Alt text on all images
- ✓ No broken links (404s)
- ✓ Core Web Vitals in "Good" range (LCP < 2.5s, FID < 100ms, CLS < 0.1)
- ✓ Mobile-friendly and responsive design
- ✓ XML sitemap submitted to search engines
- ✓ Important pages within 3 clicks of homepage
Ready to optimize your website?
Analyze Your Site Now