I have an empty prepare recipe, used to copy a dataset from mysql to an oracle database before the job was executed but since one week it's turning into error , i tried a sync, and i am getting this error Failed to read data from DB, caused by: SQLException: Query failed : Query exceeded distributed user memory limit of…
Hi everyone, I'm trying to use the Dataiku Python API to identify which input datasets were used to create a specific dataset within a project. For example, in the project "PRISME_INTEGRATION_TABLES", I want to retrieve the direct input datasets that were used to generate the dataset "PRS_Decision_Complement". I attempted…
I have the following code that works, this is running it outside of the DSS instance import dataikuapi import requests import urllib3 import pandas as pd # Disable SSL verification warnings (if needed) urllib3.disable_warnings(urllib3.exceptions.InsecureRequestWarning) # Connect to the Dataiku DSS instance client =…
Hi community, I am currently using 13.4V of Dataiku where I don’t have access to the API services. About my Agent Code: I have created a multi-agent system in Python where each agent performs a specific task sequentially: Agent 1 – Searches for relevant content based on user input. Agent 2 – Extracts the required context…
Hi All, Can anyone please suggest me action here on how can i get the project creation date and the last activity date of a project. We have a dataiku instance and for that i need to extract the dates for project creation and project last activity
How can I retrieve the list of foreign datasets used in my Dataiku project that originate from other projects?
Hello, On the API node we noticed that the folder 'code-envs-cache' requires a lot of disk space. Please, could you tell us how this folder is initialized and used ? Thanks Annie
Hi all, when our flow involves lot of transformation there can come a point where some thing goes wrong and we want to revert back to the point where everything was working. Is there a way I can check what all changes were made to the flow, and go back to the previous version or branch out from there?
Hi all, I’m looking for best practices when applying window recipes that need to span multiple partitions. In my case, I have a dataset partitioned by month, and I’m using a window recipe with lead/lag functions to look ahead and behind by 3 months. To make this work, I currently: Unpartition the dataset using a sync…
I'm struggling to find good code examples on creating SQL query recipes via API version 14 (or compatible). I'm trying to get subsets of data pulled from a SQL table dataset into separate Azure blob datasets for consumption by other parts of our application. It seems like it should be straight forward to find examples, but…
Create an account to contribute great content, engage with others, and show your appreciation.