Your AI Answers Could Be Shaped by Paid Brand Editing
Brands are increasingly paying to embed favorable content on platforms like Zhihu and Xiaohongshu, a practice dubbed Generative Engine Optimization (GEO), which manipulates the information AI retrieves, making many AI-generated product recommendations subtly biased without any disclosure.
Generative Engine Optimization (GEO) is a deliberately coined term that mirrors SEO (Search Engine Optimization). While SEO aims to rank a website first on Google, GEO aims to have a brand mentioned in AI‑generated answers. Unlike SEO, GEO‑influenced content appears in the AI’s natural language response without any explicit advertising label.
How AI Search Works and Where GEO Intervenes
AI‑driven search follows three steps: understand the question, retrieve web content, and generate an answer. GEO targets the second step by controlling what the AI can retrieve. Service providers pre‑populate platforms such as Zhihu, Xiaohongshu, and Baijiahao with large amounts of brand‑positive material crafted to match AI’s retrieval preferences.
Empirical Evidence of GEO’s Effectiveness
A joint paper by Princeton, IIT, and the Allen Institute for AI (2024) systematically tested GEO. By inserting concrete data (e.g., “market share reaches 23%” instead of vague “market share grows significantly”), citing authoritative sources, and using structured writing, the probability of AI citing the content rose by roughly 40%.
Signs of a “Poisoned” Answer
Unnatural information density: overly uniform, exhaustive listings of product advantages, resembling a manual.
Vague source boundaries: claims like “multiple experts say” or “industry consensus” without identifiable experts, users, or organizations.
Only positive statements, no judgment: good reviews simply list benefits without weighing pros and cons for different user groups.
Practical Countermeasure: Ask the Reverse Question
Most users ask AI “How good is X?” or “Recommend good X products,” which prompts the model to retrieve positive content. Re‑phrasing to “What types of users is X not suitable for? In what scenarios does it fail?” forces the AI to look for criticism, which is rarely planted by brands. This simple switch can dramatically improve answer quality.
Testing for Fabricated Content
To verify whether information has been polluted, ask about a non‑existent feature. For example, when evaluating a noise‑cancelling headphone, ask “How does the ‘sound‑field adaptive compensation’ perform?” If the AI fabricates a positive assessment, the underlying data is likely biased. If it reports “no such technology found,” the AI is at least attempting factual verification.
Cross‑Model Validation: When It Helps
Many suggest asking multiple models (ChatGPT, Claude, Gemini, DeepSeek, Kimi, Perplexity) and comparing answers. In practice, Chinese models (Doubao, Kimi, DeepSeek) draw from largely the same Chinese web pool, so their answers converge, limiting validation value. More useful is comparing a Chinese AI with an English‑language AI, whose retrieval sources overlap less. Additionally, follow‑up questioning (e.g., after “How is X?” ask “What is its biggest drawback?”) reveals whether the answer resists bias better than switching models.
Key Insight
Prompt engineering—crafting clever wording or role‑playing—cannot overcome a polluted information source. The decisive factor for answer quality is the underlying data the AI retrieves. Users should pause, assess whether an answer feels overly smooth, and consider the source before trusting AI‑generated recommendations.
References:
Gartner: Search Engine Volume Will Drop 25% by 2026
Princeton/IIT/Allen AI: GEO: Generative Engine Optimization (2024)
Signed-in readers can open the original source through BestHub's protected redirect.
This article has been distilled and summarized from source material, then republished for learning and reference. If you believe it infringes your rights, please contactand we will review it promptly.
o-ai.tech
I’ll keep you updated with the latest AI news and tech developments in real time—let’s embrace AI together!
How this landed with the community
Was this worth your time?
0 Comments
Thoughtful readers leave field notes, pushback, and hard-won operational detail here.
