There are two types of SEO audits. The first is an automated export from Ahrefs or Semrush, slapped with an agency logo, telling you that you have 42 missing meta descriptions. This is useless.
The second is a forensic engineering audit. It requires inspecting server logs, analyzing the rendered Document Object Model (DOM), measuring thread-blocking scripts, and mapping the true crawl architecture of the site.
Before WebMarv writes a single line of content or builds a new conversion funnel, we run our 23-point Technical Baseline Audit. Because if the foundation is broken, nothing else matters.
Here are the 5 most critical phases of that checklist.
Phase 1: The Render Pipeline (JavaScript SEO)
Googlebot is getting better at executing JavaScript, but it is still lazy and resource-constrained. If your site relies entirely on Client-Side Rendering (CSR) — looking at you, standard React and Vue apps — you are invisible until Google decides to spend the computing power to render you.
- Raw HTML vs Rendered HTML: We compare the source code to the rendered DOM. If the core content or internal links only exist in the rendered DOM, we mandate a shift to Server-Side Rendering (SSR) or Static Generation (SSG).
- Hydration Bottlenecks: If the HTML loads fast but the page is frozen while massive JS bundles execute, we identify the specific components causing the main thread block.
- Dynamic Schema: We verify that JSON-LD structured data is present in the initial HTML payload, not injected via JS three seconds later.
Phase 2: Crawl Architecture & Budget
Enterprise sites, especially ecommerce platforms, frequently accidentally create infinite numbers of URLs.
- Faceted Navigation Traps: E-commerce filters (size, color, brand) often create unique URLs that Google crawls endlessly. We map the parameter handling and enforce strict
rel="canonical"rules ornoindextags for non-valuable filter combinations. - Log File Analysis: We don't guess what Google is crawling. We pull your server logs to see exactly where Googlebot is spending its time. If 80% of crawls are hitting outdated API endpoints or paginated category pages, we restructure the internal linking.
- XML Sitemap Integrity: A sitemap should only contain 200 OK, canonical, indexable pages. We scrub sitemaps of redirects and errors that degrade Google's trust in the map.
Phase 3: Server & Edge Performance
Speed is a visibility factor and a conversion factor.
- Time to First Byte (TTFB): If your server takes 800ms just to start sending data, your caching architecture is broken. We audit CDN configurations, edge caching rules, and database query efficiency.
- Asset Delivery: We check for modern formats (WebP/AVIF), proper
srcsetimplementation for responsive images, and preloading of critical fonts and CSS. - Protocol Audits: Ensuring the server is utilizing HTTP/2 or HTTP/3 for multiplexed asset downloading.
Phase 4: Semantic DOM Hierarchy
This is crucial for Answer Engine Optimization (AEO). AI models rely on strict HTML semantics to understand the relationship between facts.
- Heading Logic: Not just checking for an H1, but ensuring the H2s and H3s create a logical, nested outline of the topic without skipping levels.
- Semantic Tags: Replacing meaningless
<div>wrappers with<article>,<aside>,<nav>, and<main>to guide machine parsers. - Table and List Formatting: Ensuring tabular data uses proper
<th>and<td>tags, which AI engines heavily favor for extracting facts.
Phase 5: Security & Status Codes
- Redirect Chains: Finding links that go through 3 or 4 redirects before hitting the destination. We flatten these to single hops to preserve link equity.
- Mixed Content: Ensuring absolutely zero HTTP assets load on HTTPS pages.
- Soft 404s: Identifying pages that say "Not Found" but return a 200 OK status code, confusing search engines.
A beautiful website built on bad architecture is like a mansion built on a swamp. Let the engineers inspect the foundation.


