1. What is crawl budget and why does it matter? [Junior]
What the interviewer wants: Do you understand how Googlebot allocates its resources?
Strong answer: Crawl budget is the number of pages Googlebot will crawl on your site within a given timeframe. It’s determined by two factors: crawl rate limit (how fast Google can crawl without overloading your server) and crawl demand (how much Google wants to crawl based on popularity and freshness). For most sites under 10,000 pages, crawl budget isn’t a concern. It matters for large sites (100K+ pages) where inefficient crawling means important pages don’t get indexed. You manage it by fixing crawl traps, removing low-value pages from the index, and keeping your XML sitemap clean.
2. Explain the difference between indexing and crawling. [Junior]
What the interviewer wants: Can you distinguish the two stages of search engine discovery?
Strong answer: Crawling is when Googlebot visits a URL and downloads its content. Indexing is when Google processes that content, understands it, and stores it in its search index. A page can be crawled but not indexed (Google visited but chose not to include it). Common reasons for crawled-not-indexed: thin content, duplicate content, noindex tag, or the page didn’t meet quality thresholds. You can verify indexing status in Google Search Console’s URL Inspection tool.
3. What are Core Web Vitals and how do you improve them? [Junior]
What the interviewer wants: Do you know the current page experience metrics?
Strong answer: Core Web Vitals measure loading performance (LCP), interactivity (INP, which replaced FID in March 2024), and visual stability (CLS). LCP should be under 2.5 seconds. INP should be under 200 milliseconds. CLS should be under 0.1. To improve LCP: optimize images, implement lazy loading, upgrade hosting, and minimize render-blocking resources. For INP: reduce JavaScript execution time, break up long tasks, and optimize event handlers. For CLS: set explicit dimensions on images/ads and avoid dynamically injected content above the fold.
4. How does JavaScript rendering affect SEO? [Mid]
What the interviewer wants: Can you handle modern, JS-heavy websites?
Strong answer: Google uses a two-phase process: it crawls the HTML first, then queues the page for rendering with its Web Rendering Service (WRS). There can be a delay between crawl and render, meaning content loaded only via JavaScript may not be indexed immediately. Client-side rendered (CSR) frameworks like React or Vue are riskier for SEO than server-side rendering (SSR) or static site generation (SSG). Best practices: use SSR or pre-rendering for critical content, ensure internal links are in the initial HTML, and test with Google’s Rich Results Test or URL Inspection tool to verify what Google sees after rendering.
5. What is a canonical tag and when would you NOT use one? [Mid]
What the interviewer wants: Do you understand the nuances, not just the definition?
Strong answer: A canonical tag tells Google which version of a page is the “primary” version when duplicate or near-duplicate content exists. You’d use it for print pages, URL parameter variations, or syndicated content. When NOT to use one: don’t canonical paginated pages to page 1 (use rel=”next/prev” or let Google handle it). Don’t canonical pages with substantially different content. Don’t use canonical across different domains unless the content is truly identical. And never use canonical as a substitute for 301 redirects when pages have permanently moved.
6. How would you handle a website migration from HTTP to HTTPS? [Mid]
What the interviewer wants: Can you plan a project that affects the entire site?
Strong answer: Start by auditing all existing URLs. Implement the SSL certificate across all subdomains. Set up 301 redirects from every HTTP URL to its HTTPS equivalent. Update all internal links to HTTPS. Update the sitemap to HTTPS URLs. Update Google Search Console (add the HTTPS property). Update canonical tags. Check for mixed content (HTTP resources loaded on HTTPS pages). Monitor Google Search Console for crawl errors for 30-60 days post-migration. Expect a 2-4 week ranking fluctuation. Run the migration during a low-traffic period.
7. What is hreflang and when is it needed? [Mid]
What the interviewer wants: International SEO knowledge.
Strong answer: Hreflang is an HTML attribute that tells Google which language and regional version of a page to serve to users in different locations. It’s needed when you have the same content in multiple languages (en, fr, de) or the same language targeting different regions (en-US, en-GB, en-AU). Implementation options: HTML link tags, HTTP headers (for PDFs), or XML sitemap entries. Common mistakes: missing return tags (if page A points to page B, page B must point back to page A), incorrect language codes, and forgetting the x-default for fallback.
8. Explain how Google handles duplicate content. [Junior]
What the interviewer wants: Do you understand the reality vs. the myths?
Strong answer: Google doesn’t penalize duplicate content in most cases. It simply chooses one version to index and ignores the others. Google picks the “canonical” version based on signals like which URL has more backlinks, which is in the sitemap, and which canonical tag points where. The real problem with duplicate content isn’t penalties, it’s diluted signals: backlinks and engagement split across multiple URLs instead of consolidating on one. Fix it with canonical tags, 301 redirects, or parameter handling in Google Search Console.
9. What is structured data and how does it affect rankings? [Junior]
What the interviewer wants: Do you know the difference between rankings and visibility?
Strong answer: Structured data (schema markup) is code in JSON-LD, Microdata, or RDFa format that helps search engines understand page content. It doesn’t directly improve rankings (Google has confirmed this), but it enables rich results: star ratings, FAQs, how-to steps, product pricing, event dates, and more. Rich results increase click-through rates by 20-30% (Search Engine Journal, 2024), which indirectly helps performance. Common schemas: FAQPage, HowTo, Article, Product, LocalBusiness, and Organization.
10. How do you diagnose and fix a sudden drop in organic traffic? [Senior]
What the interviewer wants: Systematic problem-solving under pressure.
Strong answer: Follow a diagnostic tree. First, check if it’s site-wide or page-specific (Google Search Console > Performance > Pages). If site-wide: check for a manual action (Search Console > Security & Manual Actions). Check for a Google algorithm update (Search Engine Roundtable, SERP tracker tools). Check server logs for increased crawl errors. Check for accidental noindex tags or robots.txt blocks (a deployment that overwrites robots.txt is common). If page-specific: check if a competitor launched better content. Check if the page’s content was changed. Check if backlinks were lost. Check search intent shift. Timeline matters: correlate the drop date with any deployments, algorithm updates, or Google Search Console messages.
11. What’s the difference between 301 and 302 redirects? When would you use each? [Junior]
What the interviewer wants: Basic redirect knowledge with practical application.
Strong answer: A 301 is a permanent redirect that passes 95-99% of link equity to the destination URL. Use it for permanent URL changes, site migrations, and consolidating duplicate pages. A 302 is a temporary redirect that tells Google to keep the original URL indexed. Use it for A/B testing, temporary maintenance pages, or geo-redirects. Common mistake: using 302s when you mean 301s, which prevents link equity transfer and can cause indexing confusion.
12. How do you optimize a site’s XML sitemap? [Mid]
What the interviewer wants: Do you know what belongs in a sitemap and what doesn’t?
Strong answer: A sitemap should only include URLs you want indexed. Remove 404 pages, redirected URLs, noindexed pages, paginated pages (usually), and thin/duplicate content. Include lastmod dates (but only update them when content actually changes, not on every crawl). Keep each sitemap file under 50,000 URLs or 50MB. Use sitemap index files for larger sites. Submit via Google Search Console and reference in robots.txt. Monitor the sitemap coverage report for issues.
13. What is log file analysis and what can it tell you? [Senior]
What the interviewer wants: Advanced technical debugging ability.
Strong answer: Log file analysis examines your server’s access logs to see exactly which URLs Googlebot (and other bots) are requesting, how often, and what response codes they receive. It reveals: which pages Google crawls most frequently, which pages Google never visits, crawl errors that Search Console doesn’t report, crawl budget waste on non-essential pages, and whether Google is rendering JavaScript pages. Tools: Screaming Frog Log Analyzer, Botify, or custom scripts parsing Apache/Nginx logs. It’s especially valuable for sites with 100K+ pages where Search Console data alone is insufficient.
14. How do you handle faceted navigation for SEO? [Senior]
What the interviewer wants: Ecommerce SEO depth.
Strong answer: Faceted navigation (filters for color, size, price, brand on ecommerce sites) can create thousands of URL combinations that waste crawl budget and create duplicate content. The strategy depends on scale. For small catalogs: noindex filter pages, canonical to the main category. For large catalogs: use robots.txt to block filter parameter URLs from crawling, implement AJAX-based filtering that doesn’t create new URLs, or use Google’s URL Parameters tool (though it’s been deprecated, the principle remains). The key decision: which filter combinations have enough search volume to justify their own indexable page? “Red Nike running shoes” might deserve a page. “Red size-9 Nike running shoes on sale” probably doesn’t.
15. What is E-E-A-T and how do you optimize for it? [Mid]
What the interviewer wants: Understanding of quality signals beyond on-page factors.
Strong answer: E-E-A-T stands for Experience, Expertise, Authoritativeness, and Trustworthiness. It’s not a ranking factor or score but a framework Google’s quality raters use to evaluate content quality. You can’t directly optimize for it, but you can signal it: author bios with credentials, bylines on all content, external mentions and citations, clear about pages, transparent contact information, accurate factual claims with sources, and first-hand experience demonstrated in content. E-E-A-T matters most for YMYL (Your Money, Your Life) topics like health, finance, and legal content.