Analysis of Parameter Count Estimates for Frontier AI Models: Grok 3, Claude 3.7 Sonnet, and GPT-4.5
Recent advancements in large language models (LLMs) have sparked intense interest in understanding their architectural scale, particularly regarding parameter counts. This analysis synthesizes available evidence to estimate the parameter sizes of three frontier models: xAI's Grok 3, Anthropic's Claude 3.7 Sonnet, and OpenAI's GPT-4.5. While manufacturers typically withhold exact architectural details, multiple independent analyses and technical comparisons provide credible estimates.
xAI's marketing claims about Grok 3 having "10x the computational power" of Grok 2[^4] require careful interpretation. The 100,000 Nvidia H100 GPUs mentioned in implementation details[^4] suggest substantial distributed training infrastructure rather than direct parameter scaling. Computational power metrics (FLOPs) correlate with but don