-
mise à jour hebdomadaire d'un fichier csv uploadé
Bonjour, chaque semaine, j'ai un scénario qui prend en compte un fichier csv uploadé, également chaque semaine. J'au donc un dataset uploadé qui s'appelle contrat_fraude_verif sur lequel commence un flow (voir ci-dessous). Quelle est la meilleure pratique pour avoir un minimum d'opérations à faire lors de l'intégration…
-
DSS engine error : while joining Database Engine and Spark Configuration Issue
When I attempt to join datasets using the DSS engine, I encounter an error stating that the recipe cannot utilize the in-database engine and will default to the slower DSS engine instead. Additionally, it warns that the 'national_joined' dataset is not a SQL table dataset. If I switch to the Spark engine, I receive a…
-
How can I use the as_langchain_chat_model() method to get the langchain compatible llm?
Hello, I am a new user of LLM Mesh and I am experimenting with creating a multi-agent workflow within Dataiku. The documentation specifies that there is a way to get a langchain compabible llm by using the as_langchain_chat_model() method (https://developer.dataiku.com/latest/api-reference/python/llm-mesh.html). However,…
-
dataiku failed to load data from snowflake(large dataset)
Exception: Reading dataset failed: failed to read data from table, caused by: SnowfalkeSQLLoggedException: JDBC driver internal error: Timeout waiting for the download of #chunk0(Total chunks: 79)
-
I'm somewhat embarrassed as this should be basic and incredibly simple but ...
I'm working with CSV files and developed a couple of recipes with one dataset until they did what I wanted. I assumed that replacing the development dataset would be dead simple and I would just click to replace the old data set and be able to specify the new csv file and run the recipes with the new data using the UI. The…
-
Create dataset based on criteria
I have a folder of 40k+ images, and I have created a separate dataset from multiple sources that resulted in a list of image names that I want to keep, plus some key attributes. How can I create a new image folder from my big image folder that only contains the images with names that match a column in one of my prepared…
-
Test Amazon S3 connection
I setup a Amazon S3 connection to a Minio server using an http endpoint and using path style option. When i click on the TEST button , the following error appears: Connection failed: Bad Request (Service: Amazon S3; Status Code: 400; Error Code: 400 Bad Request; Request ID: 17EC370F655CA6AA; S3 Extended Request ID:…
-
Question about the install path of Dataiku
Due to the space in my username(Zhao Guanghao)of my laptop, I can not run dataiku correctly. Meanwhile, the filename under the 'User' file in C disk can't be modified, once you rename it directly, the system can not identify the account and it will collapse. Besides, Dataiku can only be installed in the path 'C…
-
dataiku internal client not compatible with Python 3.12
Hi Dataiku Guru, We used the dataiku internal client outside of Dataiku in a python script, but in python 3.12, when import dataiku module an error raised due to the depreciation of imp python module in pyhon 3.12 https://github.com/python/cpython/issues/98040 The dataiku/base/utils.py still uses the imp When will the…
-
Creating a data extract with a SQL and passing a filter parameter
I am trying to create a data extract from SnowFlake . I have selected the connection and written my sql . I want this SQL to be available to multiple users such that they can apply filters to the dataset to retrieve the information . For example, in my dataset I have a "COUNTRY" column. I want a way for users to use the…