Hello Dataiku, I am trying to install packages on a new kernel that I have built using the "python env" code environment option in the administration page. However when I try to install the package, it still clashes with the default python's packages on the machine. This is also evident when I execute `!pip list` and check…
Bonjour, chaque semaine, j'ai un scénario qui prend en compte un fichier csv uploadé, également chaque semaine. J'au donc un dataset uploadé qui s'appelle contrat_fraude_verif sur lequel commence un flow (voir ci-dessous). Quelle est la meilleure pratique pour avoir un minimum d'opérations à faire lors de l'intégration…
When I attempt to join datasets using the DSS engine, I encounter an error stating that the recipe cannot utilize the in-database engine and will default to the slower DSS engine instead. Additionally, it warns that the 'national_joined' dataset is not a SQL table dataset. If I switch to the Spark engine, I receive a…
Hello, I am a new user of LLM Mesh and I am experimenting with creating a multi-agent workflow within Dataiku. The documentation specifies that there is a way to get a langchain compabible llm by using the as_langchain_chat_model() method (https://developer.dataiku.com/latest/api-reference/python/llm-mesh.html). However,…
I'm working with CSV files and developed a couple of recipes with one dataset until they did what I wanted. I assumed that replacing the development dataset would be dead simple and I would just click to replace the old data set and be able to specify the new csv file and run the recipes with the new data using the UI. The…
I have a folder of 40k+ images, and I have created a separate dataset from multiple sources that resulted in a list of image names that I want to keep, plus some key attributes. How can I create a new image folder from my big image folder that only contains the images with names that match a column in one of my prepared…
I setup a Amazon S3 connection to a Minio server using an http endpoint and using path style option. When i click on the TEST button , the following error appears: Connection failed: Bad Request (Service: Amazon S3; Status Code: 400; Error Code: 400 Bad Request; Request ID: 17EC370F655CA6AA; S3 Extended Request ID:…
Hi Dataiku Guru, We used the dataiku internal client outside of Dataiku in a python script, but in python 3.12, when import dataiku module an error raised due to the depreciation of imp python module in pyhon 3.12 https://github.com/python/cpython/issues/98040 The dataiku/base/utils.py still uses the imp When will the…
I am trying to create a data extract from SnowFlake . I have selected the connection and written my sql . I want this SQL to be available to multiple users such that they can apply filters to the dataset to retrieve the information . For example, in my dataset I have a "COUNTRY" column. I want a way for users to use the…
I have uploaded a CSV and stored it in the filesystem_folders. I have built several recipes from this dataset. I have now received an updated version of the CSV, but cannot figure out how to upload it and overwrite the original dataset. It seems to require I create a new dataset. If I do create a new dataset, there doesn't…
Create an account to contribute great content, engage with others, and show your appreciation.