/
Navigation
C
Chronicles
Browse all articles
C
E
Explore
Semantic exploration
E
R
Research
Entity momentum
R
N
Nexus
Correlations & relationships
N
~
Story Arc
Topic evolution
S
Drift Map
Semantic trajectory animation
D
P
Posts
Analysis & commentary
P
Browse
@
Entities
Companies, people, products, technologies
Domains
Browse by publication source
Handles
Browse by social media handle
Detection
?
Concept Search
Semantic similarity search
!
High Impact Stories
Top coverage by position
+
Sentiment Analysis
Positive/negative coverage
*
Anomaly Detection
Unusual coverage patterns
Analysis
vs
Rivalry Report
Compare two entities head-to-head
/\
Semantic Pivots
Narrative discontinuities
!!
Crisis Response
Event recovery patterns
Connected
Nav: C E R N
Search: /
Command: ⌘K
Embeddings: large
Entity

smaller models

5 articles stable
Articles
5
mentions
Velocity
0.0%
growth rate
Acceleration
+0.667
velocity change
Sources
4
publications

Coverage Timeline

2025-11-18
@artificialanlys

Artificial Analysis announces AA-Omniscience, a benchmark for knowledge and hallucination across 40+ topics; Claude 4.1 Opus takes first place in its key metric

@artificialanlys : X: @artificialanlys , @emollick , @scaling01 , @teortaxestex , @artificialanlys , @zephyr_z9 , @artificialanlys , @artificialanlys , @mweinbach , @artificialanlys , and @artificial...

2024-12-15
Nature 4 related

AI companies, running out of conventional training datasets from the web, may be forced to shift from big, all-purpose LLMs to smaller, more specialized models

why human-sourced data can help prevent AI model collapse Matthias Bastian / The Decoder : OpenAI co-founder says AI is reaching “peak data” as it hits the limits of the internet Kylie Robison / The V...

2024-08-24
Wall Street Journal 4 related

Chinese AI startups, cut off from the most powerful AI chips, are focusing on monetization, writing more efficient code for LLMs, and building smaller models

Wall Street Journal :

2024-08-23
Wall Street Journal 1 related

Chinese AI startups, cut off from the most powerful AI chips, are focusing on monetization, writing more efficient code for LLMs, and building smaller models

Wall Street Journal :

2024-07-25
Financial Times 9 related

Researchers suggest that using “synthetic” data, created by AI systems to train LLMs, could lead to the rapid degradation of AI models and a collapse over time

In April this year, OpenAI CEO Sam Altman was asked during … Lindsay Clark / The Register : AI models face collapse if they overdose on their own output Tech Xplore : Using AI to train AI: Model colla...

Loading articles...

Quarterly Coverage

Top Sources

Narrative

Loading narrative...

Relationships

Loading graph...