スキップしてコンテンツを表示

Content correction for aws.amazon.com/what-is/foundation-models "GPT-4 using 170 trillion parameters and a 45 GB training dataset."?

0

Hi there,

On the following page: https://aws.amazon.com/what-is/foundation-models/

It says: “In 2023, only five years later, OpenAI trained GPT-4 using 170 trillion parameters and a 45 GB training dataset.”

I believe this may be a mistake:

  • 170 trillion parameters seems unrealistic compared with publicly discussed estimates (around 1.7–1.8 trillion), and OpenAI has not disclosed GPT-4’s parameter count in its official technical report.
  • 45 GB training dataset appears far too small for a model of GPT-4’s scale. On another AWS page, “What is GPT?”, you mention 175 billion parameters and ~45 TB of data for GPT-3, and provide no dataset size for GPT-4.

Could you please review and confirm whether these figures are accurate?

Kind regards, Marshall

質問済み 3ヶ月前49ビュー
1回答
0

You're correct to question these figures. The statement about GPT-4 having "170 trillion parameters and a 45 GB training dataset" appears to be inaccurate based on available information.

OpenAI has not publicly disclosed the exact parameter count for GPT-4 in their technical report. While estimates exist in the industry, the 170 trillion figure is significantly higher than commonly cited estimates, which tend to be in the range of 1-2 trillion parameters.

Regarding the training dataset size, 45 GB would indeed be extremely small for a model of GPT-4's scale and capabilities. For context, GPT-3 (a predecessor model) was trained on a much larger dataset measured in terabytes, not gigabytes.

These figures should be reviewed for accuracy, as they appear to contain errors that could mislead readers about the technical specifications of GPT-4.
Sources
Community | A Network Engineers Guide to Generative AI

回答済み 3ヶ月前

ログインしていません。 ログイン 回答を投稿する。

優れた回答とは、質問に明確に答え、建設的なフィードバックを提供し、質問者の専門分野におけるスキルの向上を促すものです。

関連するコンテンツ