-
Retrain the Partition Model
Hi, Initially i have trained the model for partition dataset and deployed the partitioned model (Partitioned on CITY column) . with model Id = 'XYZ' The requirement is to re train the partitioned model for all the partitions (The Number of partitions changes every time the dataset is reloaded) and activate the latest…
-
How to get regression coefficients of each model build on a partitioned dataset with python code?
Hi , I have build and deployed liner regression model on a partitioned data, so there is a regression model for each partition. I want to get the regression coefficients of each of those model with a python code or recipe (basically to automate, I do not want to download the coefficients manually). Does anyone has any idea…
-
How to create a hierarchy
Hello, I would like some help to create a hierarchy. Indeed, I have some issues to create one. I have a a raw file such the way explained on the raw excel data (colored) with parents and child IDs and Names and I would like to be able to get 3 tables : One with all the parent and child name, one with all their ID and one…
-
Saving an image in the box folder connected to datiku
I have a screenshot function as part of my Rshiny app. Currently, I'm saving the screenshot as follows in a datiku managed folder. This code is working fine, and I'm able to save the image as a PNG file screenshot_folder <- dkuManagedFolderPath("SCREENSHOT_FOLDER") screenshot_path <- paste0(screenshot_folder,"/")…
-
comment faire une historisation des données par mois d'un tableau d'un autre projet .
-
how to increase character length for a column in a table I am loading from Oracle to Snowflake
I keep getting the below error when I am trying to load data from oracle to snoowflake using a dataiku sql query recipe. User character length limit (25) exceeded by string 'Gomez Hidalgo del Castillo' File 'snowflake_stage/snowflake_stage/tmp.FBe55Q5TgEnla6Jj/out-s0-c1.csv.gz', line 2939, character 1 Row 2939, column…
-
Using the API, what are my options for finding the last time a scenario ran in a project?
I have tried looping through the project's scenarios using get_last_finished_run(), but it looks as if that throws an exception if a scenario has never run, which can be the case. Since that loop is already within a try/except block, the coding gets tricky. Any help much appreciated here. Operating system used: AWS
-
Simplest way to get the aggregate value from one dataset, and bring it in to another
I have dataset A and dataset B. I need the aggregate total from one column called "Total Commission" from B. I want to bring it into A and populate a single column with that value. I know I can do this in Python with two dataframes and I know I can do this with a join if I create a join key in the datasets. Is there a…
-
Data validation compared to previous data
Hello, is there a way to check and validate data? I have webshop traffic data in my spreadsheet on a daily basis. These are divided into our different channels like SEA, Price Search Engines, SEO and so on. I'm looking for a way to check if there are major discrepancies in new data compared to previous ones. In this way I…
-
Handling Project variables in Scenarios
Hi Team, I have created scenario with consists of 5 steps. step 1: sql query to check latest data step 2: set project variable as execution start time using now() formula step 3: Build the dataset step 4: Run Ml model step 5: set project variable as execution end time using now() formula. Here question is step 2 and step 5…