Sign up to take part
Registered users can ask their own questions, contribute to discussions, and be part of the Community!
Hi @AlexT ,
I am using join recipe and there i am joining tow datasets one of the dataset has 7M records. It is caching that dataset in memory and running for longer period of time and later I am getting out of space issue.
Kindly help me how I can resolve this issue.
Joins with large datasets are not recommended not to be done with the DSS engine. It can reqire a significant amount of temporary disk space to be available in the DATADIR.
You should SQL Engine or Spark. You can sync your datasets to SQL Database and do the join with SQL engine instead.
If that's not an option you should try:
- clean up space https://doc.dataiku.com/dss/latest/operations/disk-usage.html or
- add additional disk space to allow for the join to succeed.
If jobs will create temp files under DATADIR/jobs/PROJECTID/RECIPENAME...