Master these 31 carefully curated interview questions to ace your next Seo interview.
SEO (Search Engine Optimization) improves website visibility in organic search results, driving free, targeted traffic.
SEO involves optimizing content, technical elements, and authority signals so search engines rank your pages higher. Types: on-page (content, meta tags, headings), off-page (backlinks, social signals), technical (speed, mobile-friendliness, crawlability). Benefits: free traffic, high intent visitors, long-term ROI, builds credibility. SEO vs SEM: SEO is organic, SEM includes paid ads (PPC). Google processes 8.5 billion searches daily.
Title tags, meta descriptions, heading structure, content quality, internal linking, URL structure, image alt text, and keyword optimization.
Critical elements: (1) Title tag: 50-60 chars, primary keyword near start. (2) Meta description: 150-160 chars, compelling CTA. (3) H1: one per page, includes keyword. (4) Heading hierarchy: H1→H2→H3. (5) Content: comprehensive, unique, answers search intent. (6) Internal links: contextual, descriptive anchor text. (7) URL: short, keyword-included. (8) Images: descriptive alt text, compressed, WebP format. (9) Schema markup for rich snippets.
On-page SEO optimizes elements on your website; off-page SEO builds authority through external signals like backlinks and brand mentions.
On-page: content quality, keyword optimization, meta tags, internal linking, page speed, mobile-friendliness, schema markup. Off-page: backlink building (quality > quantity), brand mentions, social signals, local SEO (Google Business Profile), reviews. Off-page is harder to control but signals trust and authority. Domain Authority (Moz), Domain Rating (Ahrefs) measure off-page strength. E-E-A-T (Experience, Expertise, Authoritativeness, Trustworthiness) spans both.
Technical SEO ensures search engines can crawl, index, and render your website effectively through site structure and performance optimization.
Key areas: (1) Crawlability: robots.txt, XML sitemap, clean URL structure. (2) Indexability: canonical tags, noindex/nofollow, hreflang. (3) Performance: Core Web Vitals (LCP, FID, CLS). (4) Mobile: responsive design, mobile-first indexing. (5) Security: HTTPS. (6) Structure: schema markup, breadcrumbs. (7) JavaScript SEO: SSR/SSG for JS frameworks. Tools: Google Search Console, Screaming Frog, Lighthouse. Fix: crawl errors, duplicate content, broken links, redirect chains, orphan pages.
Core Web Vitals are Google's metrics for page experience: LCP (loading), INP (interactivity), and CLS (visual stability).
LCP (Largest Contentful Paint): <2.5s — measures loading performance. Optimize: server response time, render-blocking resources, image optimization. INP (Interaction to Next Paint): <200ms — replaces FID, measures responsiveness. Optimize: reduce JS execution, code splitting, web workers. CLS (Cumulative Layout Shift): <0.1 — measures visual stability. Optimize: set image dimensions, avoid dynamic content injection above fold. Tools: PageSpeed Insights, Chrome UX Report, web-vitals library. Ranking factor since 2021.
Keyword research identifies terms your audience searches for, analyzing volume, difficulty, and intent to prioritize content creation.
Process: (1) Seed keywords from business/product. (2) Expand with tools: Ahrefs, SEMrush, Google Keyword Planner. (3) Analyze: search volume, keyword difficulty, CPC. (4) Classify intent: informational, navigational, commercial, transactional. (5) Group into topic clusters. (6) Prioritize: high volume + low difficulty + high intent. Long-tail keywords: lower volume but higher conversion. Content mapping: assign keywords to pages, avoid cannibalization. Update research quarterly.
E-E-A-T (Experience, Expertise, Authoritativeness, Trustworthiness) is Google's quality guideline for evaluating content credibility.
Experience: first-hand knowledge of topic. Expertise: creator's knowledge/skill. Authoritativeness: reputation in the field. Trustworthiness: accuracy, transparency, safety. YMYL (Your Money or Your Life) topics require higher E-E-A-T: health, finance, legal. Improve: author bios with credentials, cite sources, update content regularly, earn quality backlinks, display contact info, secure site (HTTPS). Not a direct ranking factor but influences quality rater evaluations.
Use SSR/SSG for critical content, implement proper meta tags server-side, use dynamic rendering as fallback, and test with Search Console.
Challenges: search engines may not execute JS, delayed indexing, rendering budget. Solutions: (1) SSR (Next.js, Nuxt): render on server. (2) SSG: pre-render at build time. (3) Dynamic rendering: serve pre-rendered to bots. (4) Meta tags in server response (not client-side). (5) Proper internal linking (native <a> tags, not JS navigation). (6) Structured data in HTML. Testing: Google Search Console URL inspection (rendered HTML), Mobile-Friendly Test. Next.js App Router with RSC is ideal for JS SEO.
Analyze which pages/keywords lost rankings, compare against update focus, audit content quality, fix issues, and monitor recovery.
Steps: (1) Confirm timing aligns with update (Google Search Status Dashboard). (2) GSC: identify affected pages, queries losing impressions/clicks. (3) Analyze update focus: content quality (Helpful Content), links (link spam), or core (broad). (4) Audit affected pages: thin content, AI-generated, keyword stuffing? (5) Competitor analysis: who gained your lost rankings. (6) Improve E-E-A-T: add expert authors, update stale content, improve UX. (7) Remove/improve low-quality pages. (8) Monitor recovery (can take 2-6 months). (9) Diversify traffic sources.
Google crawls the web, indexes pages, then ranks them using 200+ signals including relevance, quality, authority, and user experience.
Process: (1) Crawling: Googlebot discovers pages via links and sitemaps. (2) Indexing: parse content, understand meaning (NLP/BERT/MUM). (3) Ranking: 200+ signals including content relevance, PageRank (link authority), E-E-A-T, Core Web Vitals, mobile-friendliness, HTTPS, user engagement signals. Major systems: RankBrain (ML for query understanding), BERT (natural language), Helpful Content System, Link Spam System, Reviews System. Updates: core updates every few months, specific system updates ongoing.
A sitemap is an XML file listing all important URLs for search engines to discover and crawl your website pages efficiently.
XML sitemap: lists URLs with priority, change frequency, and last modified date. Submit via Google Search Console. robots.txt: Sitemap: https://example.com/sitemap.xml. Types: URL sitemap (pages), image sitemap, video sitemap, news sitemap. Benefits: helps search engines discover pages (especially new or deep-linked), indicates important pages, shows content frequency. Auto-generation: WordPress (Yoast, Rank Math), Next.js (next-sitemap), custom scripts. Limits: 50,000 URLs or 50MB per sitemap, use sitemap index for larger sites. Dynamic sitemaps for large e-commerce with changing inventory.
robots.txt is a file at the root of your site that tells search engine crawlers which pages or sections to crawl or not crawl.
Location: https://example.com/robots.txt. Directives: User-agent (which crawler), Disallow (block path), Allow (override disallow), Crawl-delay (request frequency). Example: Disallow: /admin/ blocks admin pages. NOT a security measure — pages are still accessible, just not crawled. Common mistakes: accidentally blocking CSS/JS (prevents rendering), blocking entire site (Disallow: /). meta robots tag: more granular per-page control (noindex, nofollow). X-Robots-Tag HTTP header: for non-HTML files. Always check after site migration. Google ignores Crawl-delay directive.
Backlinks are links from external sites to yours. They signal authority and trust to search engines, heavily influencing rankings.
Quality factors: (1) Domain authority of linking site. (2) Relevance of linking page. (3) Anchor text (descriptive, relevant). (4) Position (editorial vs footer). (5) Follow vs nofollow (rel='nofollow' passes no authority). Link building strategies: (1) Create linkable content (original research, tools, infographics). (2) Guest posting on relevant sites. (3) Broken link building (find broken links, offer your content). (4) Digital PR (newsworthy content). (5) HARO (Help a Reporter Out). Avoid: buying links, link farms, excessive link exchanges — Google penalizes. Quality over quantity: one link from NYTimes > 100 from unknown blogs.
Schema markup is code added to pages that helps search engines understand content context, enabling rich results in SERPs.
Format: JSON-LD (recommended by Google), Microdata, RDFa. Common schemas: Article, Product, FAQ, HowTo, Review, Event, LocalBusiness, Organization, BreadcrumbList, VideoObject. Rich results: star ratings, price, availability, FAQ dropdowns, recipes, events. Implementation: <script type='application/ld+json'> in HTML head. Testing: Google Rich Results Test, Schema Markup Validator. Benefits: higher CTR (30%+ with rich results), better understanding by search engines, voice search optimization. FAQ schema can dominate SERP space. Product schema shows price and availability. Don't add schema for content that doesn't exist on the page.
Crawling discovers pages; indexing stores and understands content; ranking determines position in search results for queries.
Crawling: Googlebot follows links and sitemaps to discover URLs. Crawl budget: number of pages Google crawls per visit (important for large sites). Indexing: Google parses content, understands meaning (NLP), stores in index. Not all crawled pages are indexed (low quality, duplicate, noindex). Index coverage report in GSC. Ranking: algorithms evaluate indexed pages for each query — 200+ signals including relevance, authority, user experience, freshness. Process: query → retrieve candidates from index → rank by relevance → apply filters (safe search, location) → display results. You can crawl and index but never rank if content isn't relevant.
Canonical tag tells search engines which URL is the preferred version when identical or similar content exists at multiple URLs.
Implementation: <link rel='canonical' href='https://example.com/page'> in HTML head. Use cases: (1) HTTP vs HTTPS versions. (2) www vs non-www. (3) URL parameters (?sort=price vs base URL). (4) Paginated content. (5) Syndicated content on other sites (self-referencing canonical). (6) Mobile vs desktop URLs. Common mistakes: canonicalizing to irrelevant pages, chain canonicals (A→B→C, should be A→C), blocking canonicalized pages in robots.txt. Alternative: 301 redirect for permanent duplicates. hreflang + canonical for multilingual sites. Google treats canonical as a hint, not directive.
Focus on digital PR, data-driven content, topical authority, HARO/Connectively, and building genuine relationships in your niche.
Effective strategies: (1) Original research and data studies — journalists cite unique data. (2) Digital PR: create newsworthy content (surveys, reports, tools). (3) Broken link building: find 404s on relevant sites, offer your content as replacement. (4) Skyscraper technique: improve on linked content, reach out. (5) Podcast appearances: show notes include links. (6) HARO/Connectively: respond to journalist queries as expert source. (7) Resource page link building: find resource pages in niche, suggest your content. (8) Build tools/calculators that naturally attract links. Avoid: paid links, PBNs, excessive guest posting, irrelevant directories. Focus: topical authority through comprehensive coverage of your niche.
Use hreflang tags for language targeting, choose URL structure (subdirectory vs subdomain), and localize content properly.
hreflang: <link rel='alternate' hreflang='es' href='https://example.com/es/'> tells Google which language version to show. Include self-referencing and x-default. URL structures: (1) Subdirectories: example.com/es/ (easiest, single domain authority). (2) Subdomains: es.example.com (separate hosting). (3) ccTLDs: example.es (strongest geo-signal, multiple domains to manage). Content: translate AND localize (dates, currency, cultural references). Don't auto-translate. geo-targeting: Google Search Console international targeting settings. Common mistakes: missing hreflang return tags, mixing language and country codes, duplicate content without proper hreflang.
Search intent is the purpose behind a query — informational, navigational, commercial, or transactional — match content type to intent.
Types: (1) Informational: seeking knowledge ('what is SEO', 'how to bake bread') → blog posts, guides, tutorials. (2) Navigational: finding specific site ('Facebook login', 'Amazon') → homepage, landing pages. (3) Commercial: researching before purchase ('best laptops 2025', 'iPhone vs Samsung') → comparison pages, reviews. (4) Transactional: ready to buy ('buy iPhone 15', 'Nike shoes sale') → product pages, pricing. Analysis: search the keyword, see what Google ranks (SERP analysis). Content format: listicle, how-to, product page — match what's ranking. If you don't match intent, you won't rank regardless of quality.
301 redirect all URLs, update internal links, notify Google via Search Console, monitor rankings, and update backlinks.
Pre-migration: (1) Crawl current site (Screaming Frog) — URL list. (2) Benchmark rankings and traffic. (3) Plan URL mapping (old → new). (4) Set up new site, verify in GSC. Migration: (5) Implement 301 redirects for ALL URLs (page-by-page, not domain-level). (6) Update internal links to new URLs. (7) Update canonical tags, hreflang, sitemap. (8) Submit new sitemap in GSC. (9) Use Change of Address tool in GSC. Post-migration: (10) Monitor crawl errors, indexation, rankings daily. (11) Contact high-value backlink sources to update links. (12) Keep old domain and redirects active for 1+ year. Expect temporary ranking fluctuation.
Claim Google Business Profile, ensure NAP consistency, get local reviews, create local content, and build local citations.
Key actions: (1) Google Business Profile: complete all fields, categories, photos, posts, Q&A. (2) NAP consistency: Name, Address, Phone identical across all listings. (3) Reviews: actively request reviews, respond to all. (4) Local citations: directories (Yelp, Yellow Pages, industry-specific). (5) Local content: location pages, local events, community involvement. (6) Schema: LocalBusiness structured data. (7) On-page: city + keyword in title, H1, meta description. (8) Google Maps: embed on website. (9) Local link building: chamber of commerce, local news, sponsorships. (10) Mobile optimization: local searches are heavily mobile. Local pack (top 3 map results) depends heavily on proximity, relevance, and prominence.
Domain Authority (DA) predicts site ranking potential on 1-100 scale; Page Authority (PA) predicts individual page ranking potential.
DA/PA: Moz's proprietary metrics (not used by Google). Calculated from: quantity and quality of backlinks, linking root domains, total links. Logarithmic scale: going from 20→30 is easier than 70→80. Comparison: useful for competitor analysis, not absolute ranking predictor. Alternatives: DR (Domain Rating — Ahrefs), AS (Authority Score — Semrush). Improving DA: acquire quality backlinks, remove toxic links, create linkable content, improve internal linking. Don't obsess over DA — focus on ranking for target keywords and driving organic traffic. New sites start low (~1) and grow over time.
Google's HCS evaluates if content is written primarily for people vs search engines, rewarding original, helpful, first-hand experience content.
Signals: (1) Is content created for people first, not search engines? (2) Does it demonstrate first-hand experience and expertise? (3) Does it provide substantial, complete answers? (4) After reading, would someone feel satisfied? Red flags: (1) AI-generated content without human review. (2) Content written purely for SEO with no real value. (3) Summarizing other sites without adding value. (4) Clickbait titles. (5) Covering topics outside your expertise. Site-wide classifier: one section of bad content can affect entire domain. Recovery: remove/improve unhelpful content, improve remaining content quality, demonstrate E-E-A-T.
Target long-tail conversational keywords, earn featured snippets, optimize for questions, ensure fast mobile performance.
Voice search characteristics: longer queries (7+ words), question format (who, what, how), conversational tone, local intent (~40% of voice searches are local). Optimization: (1) Target question-based keywords ('how do I', 'what is the best'). (2) Write concise, direct answers (featured snippet format — 29-word sweet spot). (3) FAQ pages with natural language Q&A. (4) Schema markup (FAQ, HowTo, Speakable). (5) Local SEO optimization (near me queries). (6) Page speed (voice results load 52% faster than average). (7) Mobile-friendly. (8) Position zero (featured snippet) — voice assistants typically read the featured snippet answer.
Identify penalty type (manual or algorithmic), fix violations (remove bad links, improve content), submit reconsideration request, and monitor.
Detection: GSC Manual Actions section (manual penalty), or sudden traffic/ranking drop after known algorithm update (algorithmic). Manual penalty steps: (1) Read the penalty reason in GSC. (2) Fix: remove unnatural links (disavow file), thin content, cloaking, keyword stuffing. (3) Document all changes. (4) Submit reconsideration request with explanation. (5) Wait 2-4 weeks for review. Algorithmic: (1) Identify which update (Helpful Content, Link Spam, Core). (2) Compare affected pages against update guidelines. (3) Improve content quality, remove unhelpful pages, build better backlinks. (4) Recovery happens at next algorithm update (months). Prevention: follow Google guidelines, focus on quality, regular SEO audits.
Create valuable, search-targeted content organized in topic clusters, optimized for intent, and promoted through distribution channels.
Framework: (1) Keyword research: identify high-opportunity keywords. (2) Content mapping: assign keywords to content types (blog, guide, landing page). (3) Topic clusters: pillar page (broad topic) linked to cluster pages (subtopics). (4) Content calendar: consistent publishing schedule. (5) Content optimization: on-page SEO, intent matching, readability. (6) Update strategy: refresh outdated content (significant ranking boost). (7) Distribution: social media, email newsletter, outreach. (8) Measurement: organic traffic, rankings, conversions, engagement. Types: how-to guides, listicles, case studies, original research, tools, templates. Quality over quantity: one comprehensive guide > ten thin posts.
Analyze competitors' top keywords, backlink profiles, content gaps, technical health, and SERP features to find opportunities.
Steps: (1) Identify competitors: search your target keywords, see who ranks. (2) Keyword gap: find keywords competitors rank for that you don't (Ahrefs, Semrush). (3) Content analysis: what topics they cover, content quality, word count, format. (4) Backlink profile: referring domains, quality, anchor text distribution. (5) Technical: site speed, Core Web Vitals, mobile experience. (6) SERP features: who gets featured snippets, PAA, knowledge panels. (7) Content freshness: how often they update. Strategy: (1) Target competitor keyword gaps. (2) Create better content (10x content). (3) Replicate quality backlinks. (4) Win SERP features they have. Track competitor rankings monthly.
A content audit evaluates all site content for performance, quality, and relevance to decide: keep, update, merge, or remove.
Process: (1) Inventory: list all URLs (Screaming Frog crawl). (2) Data collection: organic traffic, rankings, backlinks, engagement per page (GA4, GSC, Ahrefs). (3) Categorize: (a) Keep — performing well. (b) Update — good topic, outdated content. (c) Merge — similar thin pages combine into one comprehensive page. (d) Remove/redirect — no traffic, no value, no backlinks. (4) Prioritize: high-impact updates first (pages ranking position 4-20 with good keywords). (5) Execute: update content, add 301 redirects for removed pages, submit updated sitemap. Results: typical 20-30% organic traffic increase. Perform annually or biannually.
Track organic traffic growth, keyword rankings, conversion rate, revenue from organic channel, and compare against SEO investment costs.
Metrics: (1) Organic traffic: GA4 acquisition report. (2) Keyword rankings: track target keywords in Ahrefs/Semrush. (3) Organic conversions: goals/events attributed to organic. (4) Revenue: e-commerce revenue from organic sessions. (5) Organic CTR: GSC impression vs click data. Calculation: ROI = (revenue from organic - SEO costs) / SEO costs × 100. SEO costs include: tools, content creation, link building, agency fees, developer time. Attribution challenge: organic traffic may assist conversions attributed to other channels. Time frame: SEO ROI typically 6-12 months. Compare: cost per acquisition (CPA) organic vs paid search.
Topical authority is demonstrating comprehensive expertise on a subject through interconnected content covering all subtopics.
Concept: Google trusts sites that thoroughly cover topics, not just individual keywords. Building: (1) Topic clusters: pillar page + supporting articles. E.g., 'React' pillar linked to 'React hooks', 'React testing', 'React performance'. (2) Comprehensive coverage: cover every subtopic, question, and angle. (3) Internal linking: interconnect all related content. (4) Expertise signals: author bios, credentials, first-hand experience. (5) Content quality: depth, accuracy, freshness. (6) Consistent publishing: regular new content in your topic area. Benefits: easier to rank for competitive keywords, Google surfaces your content for related queries. Measurement: traffic growth across topic cluster, rankings for long-tail variations.
Ready to master Seo?
Start learning with our comprehensive course and practice these questions.