-
Default content for project description
Hello DataIKU friends, I would like to enforce better documentation standard across our DSS project, one way I would like to do this is by having a "template" project description which each author completes, which follows a sec structure. is it possible to give DSS a default value for the "description" field when a new…
-
How to Parse XML Values in a Column Using Dataiku Visual Recipe (Preparation Recipe)?
Hello everyone, I am trying to parse XML values within a column of a Dataiku dataset using a visual recipe (preparation recipe) in Dataiku. For JSON values, I can use processors like "Unnest Object (Flatten JSON)" to extract and structure the data. However, I couldn't find a similar built-in processor for handling XML…
-
Issue with Dataiku Visual Recipe Failing to Save Data to SQL Database
Hello, I am currently using an Other SQL Database connection in Dataiku and utilizing Visual Recipes to save data to a table. In some cases, such as when uploading files or handling simple datasets, the table is successfully created and saved. However, there are instances where the process fails. I would like to understand…
-
Stock geometry / geopoint objects on Hive
Hello everyone, I would like to store a column of type geometry or geopoint in my HDFS dataset with the aim of later performing a geojoin recipe between a geometry column containing polygons and a geopoint column containing geopoints. When I try to store my column in geopoint format, I get the following error: And when I…
-
my pivote recipe is giving me an empty table
hi i have a project that was working fine ,but since i took this project to automation env some of my recipe in dev env and automation env are returning empty it important to know that the dev and automation env are running on the same production database (my hadoop datasource)
-
R Code Experiment Tracking
How can I do experiment tracking for a model based on R code/recipe. The documentation here: https://doc.dataiku.com/dss/latest/mlops/experiment-tracking/tracking.html only describes this scenario for python code. My DSS version is actually 13 Operating system used: macOS Operating system used: macOS
-
Object not serializable error on an if then else statement run on spark engine
I get the following error when I run an if then else error on spark. This runs fine on the local engine. I have if then else statements that run just fine in other sections of my workflow in the DSS. How should I interpret this and how do I debug? Job failed: Task not serializable, caused by:…
-
Programatically push cde images to a registry
In containerized execution I can prepare the base container image and push it to a registry so that containerized execution pods will pull the image directly from my container. However for cde image I can build it using dssadmin build-base-image —type cde But I cannot push it to a registry. I found a command that does…
-
API Designer and async function
Hello, i'm trying to run a background task when my endpoint is call. Here is a simple example of what i am trying to do. import asyncio async def run_task_background(parameters): job_uid = await run_task(parameters) #add job_uid in a table def api_py_function(): asyncio.run(run_task_background(parameters)) return…
-
How to dynamically output recent 3 days' data with partitioned dataset
Hi all, I am new to Dataiku world, I'd like to ask the right way to output data with specific time range with partitioning method. The thing I want to do is: Dynamically build the recent 3-days data from input datasets. (Use Time Range day partition) I've tested but the output seems to grow even more than I expect, and the…