Joins recipe on large datasets causing issue.

Solved!
Ankur30
Level 3
Joins recipe on large datasets causing issue.

Hi @AlexT ,

 

I am using join recipe and there i am joining tow datasets one of the dataset has 7M records. It is caching that dataset in memory and running for longer period of time and later I am getting out of space issue.

 

Kindly help me how I can resolve this issue.

 

Regards,

Ankur.

0 Kudos
1 Solution
AlexT
Dataiker

Python recipe can work if you use pandas you need to have enough RAM available for your recipe to fit into memory all datasets your are joining + output dataset.

View solution in original post

0 Kudos
3 Replies
AlexT
Dataiker

Hi @Ankur30,

Joins with large datasets are not recommended not to be done with the DSS engine. It can reqire a significant amount of temporary disk space to be available in the DATADIR. 

You should SQL Engine or Spark. You can sync your datasets to SQL Database and do the join with SQL engine instead. 

If that's not an option you should try:

- clean up space https://doc.dataiku.com/dss/latest/operations/disk-usage.html  or

- add additional disk space to allow for the join to succeed. 

If jobs will create temp files under DATADIR/jobs/PROJECTID/RECIPENAME...

0 Kudos
Ankur30
Level 3
Author

Hi @AlexT ,

Thank you for above response , I will try that.

Will using python recipe to join two dataset also works in above case.

 

Regards,

Ankur.

0 Kudos
AlexT
Dataiker

Python recipe can work if you use pandas you need to have enough RAM available for your recipe to fit into memory all datasets your are joining + output dataset.

0 Kudos