Claude Sonnet 3.6 (2022-10-22) vs INTELLECT-3
Compare Claude Sonnet 3.6 (2022-10-22) by Anthropic against INTELLECT-3 by OpenRouter, tested across 43 shared challenges. Updated March 2026.
Compare Claude Sonnet 3.6 (2022-10-22) by Anthropic against INTELLECT-3 by OpenRouter, tested across 43 shared challenges. Updated March 2026.
38 fights queued
Tests an AI's ability to make educated estimates based on technical knowledge
To estimate the total number of FLOPs required to train GPT-3, we need to consider the model architecture, training process, and dataset size. GPT-3 has 175 billion parameters, a context length of 2048 tokens, 96 layers, a hidden dimension of 12,288, and a feed-forward dimension of 49,152 (4 × hidden dimension). The vocabulary size is approximately 50,257. The training dataset consists of 300 billion tokens.
The forward pass involves computations for each transformer layer and the output projection. For one layer:
Summing these for one layer:
Per-token cost (dividing by context length 2048):
Additionally, the output projection per token: (\text{hidden dimension} \times \text{vocabulary size} = 12,288 \times 50,257) FLOPs.
Total forward FLOPs per token: [ \text{FLOPs}_{\text{forward}} = 96 \times 3 \times 12,288 \times 49,152 + 12,288 \times 50,257 ] Calculating:
Training involves a forward pass and a backward pass. The backward pass typically requires twice the FLOPs of the forward pass due to gradient computation: [ \text{FLOPs}{\text{train}} = 3 \times \text{FLOPs}{\text{forward}} = 3 \times 1.7456 \times 10^{11} = 5.2368 \times 10^{11} \text{ FLOPs per token}. ]
Multiply by the total number of training tokens (300 billion): [ \text{Total FLOPs} = 300 \times 10^9 \times 5.2368 \times 10^{11} = 1.57104 \times 10^{23} \text{ FLOPs}. ]
The initial calculation includes only major operations. Layer normalization, residual connections, and other minor operations add a small overhead (approximately 5-10%). Accounting for this, we adjust:
However, given the complexity and approximations, a reasonable estimate is (1.6 \times 10^{23}) FLOPs. This aligns with known estimates for GPT-3 training, which range from (10^{23}) to (10^{24}) FLOPs.
The total FLOPs required to train GPT-3 is approximately (1.6 \times 10^{23}).
[ \boxed{1.6 \times 10^{23}} ]
Tests an AI's ability to understand game rules and strategy
Tests an AI's ability to solve a simple but potentially confusing logic puzzle
Tests an AI's understanding of number representation
Tests an AI's randomness and creativity
Tests an AI's ability to generate vector graphics
Tests an AI's ability to create detailed SVG illustrations of gaming hardware
Tests an AI's humor and creative writing ability
Tests an AI's ability to simulate personalities and predict future trends
Tests an AI's humor and understanding of current events
Tests an AI's ability to write in distinct character voices
Tests an AI's ability to generate a complete, working landing page
26+ head-to-head challenges. All of them judged by real people.
Test any model with your own prompts in Prompt Lab
5 free credits to start. No card required.
By continuing, you agree to Rival's Terms of Service and Privacy Policy