내용으로 건너뛰기

Content correction for aws.amazon.com/what-is/foundation-models "GPT-4 using 170 trillion parameters and a 45 GB training dataset."?

0

Hi there,

On the following page: https://aws.amazon.com/what-is/foundation-models/

It says: “In 2023, only five years later, OpenAI trained GPT-4 using 170 trillion parameters and a 45 GB training dataset.”

I believe this may be a mistake:

  • 170 trillion parameters seems unrealistic compared with publicly discussed estimates (around 1.7–1.8 trillion), and OpenAI has not disclosed GPT-4’s parameter count in its official technical report.
  • 45 GB training dataset appears far too small for a model of GPT-4’s scale. On another AWS page, “What is GPT?”, you mention 175 billion parameters and ~45 TB of data for GPT-3, and provide no dataset size for GPT-4.

Could you please review and confirm whether these figures are accurate?

Kind regards, Marshall

질문됨 3달 전49회 조회
1개 답변
0

You're correct to question these figures. The statement about GPT-4 having "170 trillion parameters and a 45 GB training dataset" appears to be inaccurate based on available information.

OpenAI has not publicly disclosed the exact parameter count for GPT-4 in their technical report. While estimates exist in the industry, the 170 trillion figure is significantly higher than commonly cited estimates, which tend to be in the range of 1-2 trillion parameters.

Regarding the training dataset size, 45 GB would indeed be extremely small for a model of GPT-4's scale and capabilities. For context, GPT-3 (a predecessor model) was trained on a much larger dataset measured in terabytes, not gigabytes.

These figures should be reviewed for accuracy, as they appear to contain errors that could mislead readers about the technical specifications of GPT-4.
Sources
Community | A Network Engineers Guide to Generative AI

답변함 3달 전

로그인하지 않았습니다. 로그인해야 답변을 게시할 수 있습니다.

좋은 답변은 질문에 명확하게 답하고 건설적인 피드백을 제공하며 질문자의 전문적인 성장을 장려합니다.

관련 콘텐츠