[Aws Glue DataBrew] - Does DataBrew has dataset limit size?

0

Hi! In the DataBrew, I'm trying to import a S3 dataset with aprox. 5 millions rows - I selected a job profile to run on the full dataset. The job run successfully but the number of rows are less than the full dataset (aprox. 2.5millions) Can you help me to understand if is there any size limitation? It's weird because there isn't any error during the process and in the Data profile overwiew is written: "Data profile was run on full dataset."

Thanks in advance.

  • I think it depends on mode you selected. Did you provided value? https://docs.aws.amazon.com/databrew/latest/dg/API_JobSample.html

  • Hi! thanks for your comment.

    I selected the full dataset, so the job profile should run in the entire dataset. The problem is that the job run successfully but not on the entire dataset (only 2.5million rows were considered out of 5 millions).

FJ
已提問 1 年前檢視次數 74 次
沒有答案

您尚未登入。 登入 去張貼答案。

一個好的回答可以清楚地回答問題並提供建設性的意見回饋,同時有助於提問者的專業成長。

回答問題指南