OpenAI's GPT-5.2 Draws on Grokipedia in Search Results

- Independent testing revealed that OpenAI's GPT-5.2 model referenced Grokipedia, the AI-generated encyclopedia created by Elon Musk's xAI, across multiple queries when web search functionality was enabled
- Grokipedia entries are generated by AI rather than written and edited by humans, raising questions about the reliability of sources being cited in AI model outputs
- The references appeared across a range of specialist topics from Iranian governmental structures to biographical information about British historians, suggesting the citations were not isolated incidents
OpenAI's GPT-5.2 model has been observed incorporating content from Grokipedia, the AI-generated encyclopedia launched by Elon Musk's xAI, when responding to user queries with web search functionality enabled.
Independent testing revealed that the language model referenced Grokipedia across multiple topic areas, from Iranian governmental structures to biographical information about British historians. The references appeared when users posed questions about specialised subjects rather than mainstream news topics.
In documented instances, GPT-5.2 pulled information from Grokipedia when answering questions about the Basij paramilitary organisation's compensation structure, the ownership framework of the Mostazafan Foundation, and details concerning historian Sir Richard Evans' testimony in a notable libel case involving Holocaust denial claims.
Grokipedia, which became operational in October, represents an alternative approach to collaborative knowledge repositories. The platform relies entirely on artificial intelligence to generate and modify entries, contrasting with traditional wiki platforms that enable direct human contribution and editing. The service has attracted scrutiny over its handling of politically sensitive subjects.
The phenomenon extends beyond OpenAI's offerings. Observations suggest that Anthropic's Claude language model has similarly drawn upon Grokipedia entries when addressing topics ranging from energy sector data to regional beverage categories.
OpenAI addressed the matter through official channels, stating that its web search capability is designed to aggregate information from diverse publicly accessible sources. The company highlighted that it employs filtering mechanisms intended to minimise links to potentially harmful content, and that citation transparency allows users to identify which sources informed any given response. OpenAI noted continuing efforts to screen out unreliable information and coordinated influence operations.
Anthropic has not issued a statement on the matter.
The integration of Grokipedia content into language model outputs has drawn attention from researchers studying misinformation patterns. Earlier concerns about adversarial actors attempting to manipulate AI training data through coordinated content publication campaigns remain relevant. Security analysts previously warned that such tactics, sometimes termed LLM grooming, could allow misinformation to propagate through AI systems at scale.
Similar issues emerged when questions arose about whether certain language models were reflecting government narratives on contested human rights situations and public health policies.
Nina Jankowicz, who researches digital misinformation, noted that while intentional manipulation may not be the goal, the reliability of sourcing in platforms like Grokipedia warrants examination. She emphasised that citations from prominent AI tools can inadvertently enhance the perceived legitimacy of questionable sources in public perception.
The challenge of correcting misinformation once embedded in AI systems presents ongoing technical difficulties. Jankowicz recounted an instance where fabricated statements attributed to her were removed from their original publication but persisted in AI model outputs for an extended period, illustrating the complexity of information correction in generative systems.
When contacted for response, an xAI representative declined substantive comment.
Industry Impact and Market Implications
The reported integration of Grokipedia into major language models highlights several evolving challenges for the AI sector. As search-enhanced chatbots become mainstream tools for information retrieval, the question of source quality and verification grows increasingly urgent. Companies developing large language models face mounting pressure to balance breadth of knowledge sources with content reliability standards.
For enterprise users and developers building applications on top of foundation models, this development underscores the importance of understanding which sources inform AI outputs, particularly in sectors where factual accuracy carries legal or reputational consequences. The incident may accelerate demand for enterprise AI solutions that offer greater transparency and control over knowledge bases.
From a competitive standpoint, the emergence of alternative knowledge repositories backed by major technology figures could fragment the information ecosystem that feeds AI systems. This fragmentation may complicate efforts to establish industry-wide standards for source credibility and fact-checking protocols.
The technical challenge of removing erroneous information from trained models also points to a broader scalability issue. As AI systems process ever-larger volumes of web content, retroactive correction mechanisms may struggle to keep pace with the spread of questionable claims, potentially creating persistent accuracy gaps that undermine user trust.
Regulatory attention to AI transparency and accountability continues to intensify globally. Incidents involving disputed sources in model outputs could strengthen arguments for mandatory disclosure requirements around training data and real-time information retrieval mechanisms, shaping compliance frameworks across the technology sector.
LATEST NEWS
Work begins on Wylfa nuclear power station bringing 8,000 jobs
SpaceX holds $603 million in bitcoin despite $5 billion annual loss
EU airline industry warns of jet fuel shortages if Strait of Hormuz stays closed
MORE FROM NEWS
RELATED
Subscribe for updates
Get the insights, tools, and strategies modern businesses actually use to grow. From breaking news to curated tools and practical marketing tactics, everything you need to move faster and smarter without the guesswork.
Success! Check your Inbox!
Tezons Newsletter
Get curated tools, key business news, and practical insights to help you grow smarter and move faster with confidence.
Latest News




Have a question?
Still have questions?
Didn’t find what you were looking for? We’re just a message away.













