Sign up to take part
Registered users can ask their own questions, contribute to discussions, and be part of the Community!
Added on March 24, 2022 10:50AM
Likes: 0
Replies: 3
Hi @AlexT
,
I am using join recipe and there i am joining tow datasets one of the dataset has 7M records. It is caching that dataset in memory and running for longer period of time and later I am getting out of space issue.
Kindly help me how I can resolve this issue.
Regards,
Ankur.
Python recipe can work if you use pandas you need to have enough RAM available for your recipe to fit into memory all datasets your are joining + output dataset.
Hi @Ankur30
,
Joins with large datasets are not recommended not to be done with the DSS engine. It can reqire a significant amount of temporary disk space to be available in the DATADIR.
You should SQL Engine or Spark. You can sync your datasets to SQL Database and do the join with SQL engine instead.
If that's not an option you should try:
- clean up space https://doc.dataiku.com/dss/latest/operations/disk-usage.html or
- add additional disk space to allow for the join to succeed.
If jobs will create temp files under DATADIR/jobs/PROJECTID/RECIPENAME...