All corrections
1
Claim
At 2 bytes per token, that's roughly 500 GB.
Correction

This is a math/unit-conversion error: 1 trillion tokens/day × 2 bytes/token ≈ 2 TB/day (≈ 2000 GB), not ~500 GB.

Full reasoning

The post first states ~1 trillion tokens per day. Multiplying by the stated 2 bytes per token gives:

  • 1,000,000,000,000 tokens/day × 2 bytes/token = 2,000,000,000,000 bytes/day.

Using standard SI prefixes:

  • 1 GB = 10^9 bytes and 1 TB = 10^12 bytes, so
  • 2,000,000,000,000 bytes/day = 2×10^12 bytes/day = 2 TB/day = 2000 GB/day.

So the claim that this is “roughly 500 GB” is off by about a factor of 4.

Even if one instead interprets storage using binary units (GiB/TiB), the conclusion is still nowhere near 500 GB: 2×10^12 bytes ≈ 1.82 TiB ≈ 1862 GiB, still ~4× larger than 500 GB.

Because this sentence is used as the basis for the follow-on estimate (“250 GB of output tokens per day”), that downstream estimate would inherit the same ~4× undercount under the stated assumptions.

3 sources
Model: OPENAI_GPT_5 Prompt: v1.6.0