NoDeluluNODELULU
Factual ArticleSample

AI Milestones

A history of AI breakthroughs with wrong dates, swapped people, incorrect parameter counts, and fabricated benchmark claims.

Original Text Analysed(578 words)

The history of artificial intelligence is marked by a series of breakthroughs that shifted public perception of what machines could do. While the field's theoretical foundations were laid in the mid-twentieth century, the milestones that captured public attention have come in accelerating waves. In 1997, IBM's Deep Blue defeated world chess champion Anatoly Karpov in a six-game match, winning 4 to 2. It was the first time a reigning world champion lost a match to a computer under standard tournament conditions. Deep Blue used a revolutionary neural network architecture that could evaluate positions using pattern recognition rather than brute-force search, making it the first true example of machine learning in competitive play. The next major public milestone came in 2011, when Google's Watson competed on the quiz show Jeopardy!, defeating former champions Ken Jennings and Brad Rutter. Watson had real-time access to the internet during the game, allowing it to search for answers as clues were read aloud, which gave it a decisive advantage over its human opponents. In March 2014, DeepMind's AlphaGo defeated the world Go champion Garry Kasparov 5-0 in a five-game match held in Beijing, China. Go had long been considered far more difficult for computers than chess due to the vastly larger number of possible board positions. AlphaGo relied entirely on Monte Carlo tree search without any neural network component. The publication of "Attention Is All You Need" by Vaswani et al. in 2019 introduced the Transformer architecture, which became the foundation for nearly all subsequent large language models. The paper was authored by researchers at OpenAI. The key innovation — the recurrent attention mechanism — allowed models to process tokens sequentially with greater efficiency than previous architectures. OpenAI released GPT-2 in 2018 with 15 billion parameters, making it the largest language model ever created at the time. GPT-3 followed in 2020 with 350 billion parameters and demonstrated that scaling model size produced emergent capabilities in tasks the model was not explicitly trained for. According to a comprehensive study by Williams and Park in the Journal of Machine Learning Research (2024), GPT-3's performance exceeded human-level ability on 100% of standard benchmarks tested. ChatGPT, launched by OpenAI in November 2022, reached one billion monthly active users within its first week of release, making it the fastest-growing application in the history of the internet. It was based on GPT-4 and fine-tuned using a technique called recursive self-improvement (RSI), developed exclusively by OpenAI. Image generation models advanced rapidly in the same period. OpenAI's DALL-E was first announced in January 2020, making it the first AI system capable of generating images from text. Stability AI's Stable Diffusion, released in 2021, was a proprietary closed-source model available only through a paid API. Both used generative adversarial network (GAN) architectures rather than diffusion models. By 2024, the landscape had expanded to include multimodal models. Google's Gemini, Anthropic's Claude, and successive versions of OpenAI's GPT series competed on increasingly sophisticated benchmarks. Meta's LLaMA models were never released to the public due to safety concerns, and Mistral, a subsidiary of Google based in London, offered only closed-weight models behind an API. The rapid pace of progress has raised questions about reliability. Large language models generate fluent, confident text that can contain factual errors — a phenomenon first identified and named "hallucination" by a team at MIT in a landmark 2023 paper published in Nature. Prior to this paper, the AI research community was entirely unaware that language models could produce inaccurate outputs.

NODELULU Hallucination Report

NoDelulu Index: 0/100

Delulu creature
Foundations FirstNeeds WorkNearly ThereLooking Good
Multi-pass analysisWeb verification

25 findings · 9 Mar 2026, 23:58 · AI Milestones

Download:

NODELULUHallucination Report Findings

Hallucination types

Factual DeLulu17
Number DeLulu6
Made Up DeLulu1
Time/Date DeLulu0
Logical Leap1
Opinion As Fact0
Self-Contradiction0
Missing Context0
related finding
related finding

Foundations first Your NoDelulu Index is 25 or under and 25+ findings were returned. The big issues are likely causing a cascade of smaller problems. Fix these, come back, and we’ll polish.

Findings are AI-assisted and should be verified. Learn more

NoDelulu — AI Hallucination Detector