跳至內容

Content correction for aws.amazon.com/what-is/foundation-models "GPT-4 using 170 trillion parameters and a 45 GB training dataset."?

0

Hi there,

On the following page: https://aws.amazon.com/what-is/foundation-models/

It says: “In 2023, only five years later, OpenAI trained GPT-4 using 170 trillion parameters and a 45 GB training dataset.”

I believe this may be a mistake:

  • 170 trillion parameters seems unrealistic compared with publicly discussed estimates (around 1.7–1.8 trillion), and OpenAI has not disclosed GPT-4’s parameter count in its official technical report.
  • 45 GB training dataset appears far too small for a model of GPT-4’s scale. On another AWS page, “What is GPT?”, you mention 175 billion parameters and ~45 TB of data for GPT-3, and provide no dataset size for GPT-4.

Could you please review and confirm whether these figures are accurate?

Kind regards, Marshall

已提問 3 個月前檢視次數 50 次
1 個回答
0

You're correct to question these figures. The statement about GPT-4 having "170 trillion parameters and a 45 GB training dataset" appears to be inaccurate based on available information.

OpenAI has not publicly disclosed the exact parameter count for GPT-4 in their technical report. While estimates exist in the industry, the 170 trillion figure is significantly higher than commonly cited estimates, which tend to be in the range of 1-2 trillion parameters.

Regarding the training dataset size, 45 GB would indeed be extremely small for a model of GPT-4's scale and capabilities. For context, GPT-3 (a predecessor model) was trained on a much larger dataset measured in terabytes, not gigabytes.

These figures should be reviewed for accuracy, as they appear to contain errors that could mislead readers about the technical specifications of GPT-4.
Sources
Community | A Network Engineers Guide to Generative AI

已回答 3 個月前

您尚未登入。 登入 去張貼答案。

一個好的回答可以清楚地回答問題並提供建設性的意見回饋,同時有助於提問者的專業成長。