-
Merge/Group rows based on metrics
I am trying to merge/Group rows based on metric range. Present format Desired format
-
disabling Update output schemas in Python Scenario
Hi how do I disable the option that updates output schema in a Python scenario? Can't find the option in the API reference. Would be equivalent of circled below. Thanks Operating system used: Windows Operating system used: Windows
-
Ability to package environment/local variables with an API service
It would be very helpful if Dataiku allowed for packaging variables (either environment or local variables) with the capability to remap local variables as part of the deployment. Ideally there would also be an option to encrypt a variable. We have several API services that connect to other systems and require environment…
-
Integrating Dataiku with Denodo
Has anyone created an integration with Denodo? If so, did you use a JDBC connection? How did you manage user permissions? Thanks! Operating system used: Red Hat
-
snowflake connection is working fine from Dataiku , however fetching the table preview throws error
Failed to read data from table Failed to read data from table, caused by: SnowflakeSQLException: JDBC driver internal error: exception creating result java.lang.NoClassDefFoundError: Could not initialize class net.snowflake.client.jdbc.internal.apache.arrow.memory.RootAllocator at…
-
How do I send an email to the user based on a condition on count of records in a dataset?
hi, After reading the documentation, I came cross the "Compute metrics" step in Scenarios, but how do i retrieve the count of the dataset using ${stepOutput_the_metrics}? And then if the count is more than 0, I want to trigger an email to the user. I am on DSS version 13 Any help is appreciated. Thanks Operating system…
-
"Training failed" Error in Machine Learning Quick Start Project
I just installed Dataiku 13.3.3 (Free Edition) on a MacBook Pro (MacOS 15.3) using Chrome as a browser. I am going through the Machine Learning Quick Start project following instructions in Dataiku Academy. I am getting a "Training Failed" error when training the model with the error message below. I suspect the issue has…
-
How to use LLM Mesh work with LiteLLM
Hi, I'm working on an Agentic Gen AI project using the crewai package, which uses LiteLLM as the engine to connect to various Gen AI models. I would like to use Dataiku LLM Mesh, but it seems that it's not compatible with the LiteLLM. I tried to use the DKULLM and DKUChatLLM, but both of them are not working. I'm on…
-
Have a dataiku templating engine based on Python mako or jinja
Hi, Python based templating engines like jinja and mako allow users to 'print' text in various formats, using conditional logic statements like if-else and for loops. I think dataiku should offer an off the shelf Python based templating engine that would allow users to upload their template(s) and pass a `context dict` to…
-
How do Preserve chienese text format during CSV to dataiku load?
I'm using Dataiku version 13.1. I have a text dataset with around 2400 rows, mostly it's in english but around 100 rows contains chines character. My data is in csv format. I need to perform GenAI task on my dataset & load back to it to CSV Chinese characters are converting to english characters while loading data from CSV…
-
Data not writing to dataset in Python recipe
I created a Python recipe that I get data from an API. I save the response to a panda dataframe, but when I use the DataSet.write_from_dataframe to write the dataframe result my output dataset. The rows is written, but the row is empty when I check the dnb_duns_detail_json dataset. Operating system used: MacOS
-
Best Match for Fuzzy Matching
Hi, I am new to DataIku and trying to find areas of overlap in 2 datasets using fuzzy matching. Is there a way to get a numerical ranking for how close matches are, so I can identify the highest matches and remove duplicate suggestions if needed? Thanks,
-
Error in python process: <class 'ValueError'>: Numeric feature score_2 is empty
Trying to run a Recommendation system, I received this error not allowing to run the job. Received this error message, checked the database and all fields are with information. [01:16:18] [INFO] [dku.utils] - *************** Recipe code failed **************[01:16:18] [INFO] [dku.utils] - Begin Python stack[01:16:18]…
-
Issue with Python script appending data in Dataiku project
Hello, I have an issue with my Dataiku project. I wrote a Python script that appends new data from the input dataset to the output dataset. I think the problem may be related to recursion in Dataiku. Could you please suggest a solution? Thank you in advance!
-
Issues with the Google Drive plugin: file not appearing in the Managed Folder in Dataiku
Hello everyone! I’m using the Google Drive plugin in Dataiku to read and write files directly from my Google Drive account. I managed to set it up, and it works correctly when creating datasets from Drive. However, I’m having trouble creating a Managed Folder in Dataiku that points to a file in Drive. Here’s what I do:…
-
Import from PowerBI to Dataiku
Hi community! I have a strange question. Is there any way to import data from a poweri workspace into Dataiku? I can do the other way around, export a dataset into a workspace but was wondering if there was an obvious plugin I am missing. Many thanks :) *Also, is there any limitation? i.e. Gen2 Dataflow only, semantic…
-
How write the rmarkdown report created in dataiku to a managed folder?
-
How to dynamically name rmarkdown reports in scenarios?
I'm working on a project that requires me to send an R Markdown report to a Box folder via email daily. However, the emailed file has the same name each day, leading to overwrites and versioning issues (e.g., v1, v2, v3). I'd like to automate this process by adding the current date to the filename of the emailed report.…
-
How to append dataframe in existing output dataset
Hello experts, In dataiku v12.3.0, I was trying to append dataframe using write_dataframe() in existing dataset (with same schema). But it always overwrites with last dataframe even though the dataset spec is configured like: dataset.spec_item["appendMode"] = True The dataset is classified as output so it doesn't let me…
-
C compiler error while running python module 'pymc'
Hi, I am using pymc module to develop a MMM model in Dataiku. While working on it, I am experiencing this somewhat well-known error on C compiler but couldn't find any solution on Dataiku yet. Code where the error occured: import warnings import arviz as az import matplotlib.pyplot as plt import numpy as np import pandas…
-
How do I use training and prediction dataset together in Dataiku
Hi I'm using Dataiku version 13.1. I need to do text prediction using BERT for that I have training dataset. After training I need to score it using prediction dataset. I'm doing BERT using Python code recipe. Can you suggest me the steps to score the prediction dataset?
-
Changing the default font size for all new charts
I often send screenshots of the charts to my stakeholders over Teams chat to get quick confirmations/alignment on the intermediary results. A feedback I heard often is that they are having a hard time reading the legend, axis titles, values on axis etc. I don't want to manually change the font size for all these different…
-
How can I use websockets in a Webapp?
I would like to make a custom Webapp where the Python backend talks to the frontend and vice versa. I see that the Dataiku Answers webapp uses websockets and I would like to do the same. My current attempts using Flask-SocketIO did not work unfortunately, as it seems to use Werkzeug under the hood and I cannot start the…
-
How to load a mlflow model in a script to make predictions
I trained and deployed a model using mlflow in Dataiku. I want to make predictions on a test dataset using this deployed model. However, I don't want to use the "predict" visual recipe. Instead, I want to load the model in a script and make predictions. But I am not able to do it. Operating system used: Linux Operating…
-
Run Scenario after multiple scenarios have completed
Hi - I am trying to create a scenario that will auto-trigger once other time-based scenarios (in other projects) have completed. I think this is possible for 1 scenario using "Trigger after scenario" which automatically checks the status of a scenario at the frequency you set but I can't figure out how to do this using…
-
How to Sync Current Date from Greenplum to Oracle with Only the Date Part?
I have added a current_date column to my table in Greenplum using a Prepare recipe (with now() in Formula language). I want to sync this column to an Oracle database, but I need to keep only the date part of the value. For example, I want to convert a value like 2025-01-25T21:50:28.102Z into 2025-01-25 and store it as a…
-
Split / Join formula
Hi, Is it possible in a split (and only a split recipe not an sql one) with formula to check a condition by join. For example, to check that one of the child lines a of a parent b contains a certain value. If so, put all the corresponding lines a ->b* (of which at least one b satisfies the condition) in the split. Best…
-
Folding multiple columns by MULTIPLE patterns
Hello, I have a large number of columns that all follow the same pattern, with 3 parts that can change: Path#_Field_Year So either the Path #, Field, or Year could vary. For example, I could have: 1_Car_2030 1_Car_2031 1_Car_2032 2_Car_2030 2_Car_2031 2_Car_2032 Or I could have: 2_Bus_2040 2_Bus_2041 2_Bus_2042 5_Bus_2030…
-
I want Better Dataset Visualization in Dataiku Dashboard
I need to leverage Dataiku Dashboard to visualize Dataset with a better look. this is what I want (set different colors for value criteria / bold font / dataframe multindex support /etc): this is visualization of dataset currently (quite plain): I am code designer and can write python, btw, is there any solution? Operating…
-
Combining data sources from Blob Storage
Hello, I have a pretty simple problem, but somehow I'm not able to solve it. I'm using Dataiku to do ETL stuff and I need to pull the data from Blob Storage. Usually I just specify a path this is the start of my flow. But in this case, I need to create the first source dataset from multiple blob "folders" that are in a…
-
Beginner Help: Deploying an API Service with Pickle Model from Jupyter Notebook in Dataiku
Hello Dear Community, I am a complete beginner in Dataiku and have created a Jupyter Notebook as a mini test model. I used Pickle to save the model and vectorizer into a managed folder named "Models". My goal is to make this model available as an API service, but I’m struggling with the process and would greatly appreciate…
-
Je n'arrive pas à faire fonctionner cette formule pour calculer la différence entre deux dates
voici la formule: if(diff(DateDeSortie_MAJ2,val("DateEntréeDIR2S_MAJ"))>=0,"",DateDeSortie_MAJ2) merci pour votre aide
-
update DSS license - improve the documentation
Dataiku's documentation says there are 3 ways to update a DSS license but then only lists 2: """ Copy your new license file into DATA_DIR/config/license.json (needs restart of DSS) Log into Dataiku DSS, click the “Administration” gear, then click “Enter license” and enter the contents of the license file. """ I suppose the…
-
Container configuration mapping in bundle deployments
This request is to add mapping options for container configuration in bundle deployments. This would allow for repointing in the event that the container configurations are not named the same in the design node vs automation nodes.
-
Per-user credentials in LLM connections
This request is to add support for per-user credentials in LLM connections. We use OpenAI and set up API keys per project so that we can track spend and budgets at the project level. Currently we have to set up a separate OpenAI connection for each project but ideally we would be able to pass the API key in either through…
-
Selenium issues with Chrome.exe and Chromedriver.exe
I get this error Chrome binary found at: C:\\Program Files\\Google\\Chrome\\Application\\chrome.exeExecute permission for C:\\Program Files\\Google\\Chrome\\Application\\chrome.exe is granted. Failed to launch Chrome binary at C:\\Program Files\\Google\\Chrome\\Application\\chrome.exe: /bin/sh: C:\Program…
-
Bug - CTRL+Z Erases Text on Dataiku
I accidentally pressed ctrl + Z after opening a Wiki article and it caused all the text to be removed. I'm on DSS 8.0.2. This also happens within SQL queries and the Library Editor. Is this intended behavior? Here are the steps to reproduce with the Wiki: 1. Create Wiki article with some text and save 2. Navigate away from…
-
How to chart time series longer than 20 data points?
I'm creating a dashboard for reporting abnormal data series, looking at the past 90 day sales. I want to create the line chart of the past 90 day sales to monitor spikes and abnormalities. But the line chart only shows about 20 data points and lump all rest into "other" categories. Is this the default setting of chart…
-
How to get Sigmoid Platt scaling parameters used in Dataiku visual analysis Probability Calibration?
Hi Team, In the current MLOps pipeline which we are using for Model Development, we often have the need of re-calibrating the propensity so that we can have correct scaling. For this we are current using Platt Scaling which can be switched on or off from the portal as can be seen here - However , for this feature we…
-
Spark installation in standalone mode in DSS
Hi, Our Dataiku is not connected to Hadoop or Kubernetes cluster. Can I still install spark in local DSS server?
-
How do I get created and modified dates fro an artifact in Dataiku Govern API
I can see this information in the timeline in the UI, but doesn't seem to be in the data returned by: ar = client.get_artifact('ar.26421') print(ar.get_definition().get_raw()) Operating system used: Windows
-
Threshold optimization for what if? fuction in AutoMLprediction
I used "what if?" function in AutoMLprediction to do simulation, and I would like to change the threshold DSS automatically used. In the userguide, there is below description. But I could not find any button/icon to change a threshold. Please let me know how to change it. Prediction settings — Dataiku DSS 13 documentation…
-
Collaborative use of code studio template
We have recently starting working with code studio templates in order to create streamlit webapps on Dataiku. We quickly realized that while each (non admin) user can create their proper code studio instance from a specific code studio template without any issues, it is not possible for multiple (non admin) users to have…
-
Using large context for a Gen AI prompt
Hi, I'm trying to create a prompt to ask questions to a LLM and get an answer based on 5,000 reviews for a product. I know there are ways to classify or perform sentiment analysis, but what I want to do is to ask an LLM a question about the whole bunch of reviews. I tried using RAG, but it is my understanding the this…
-
Trigger on scenarios
Hello, I want to build a dataset using an append mode at the end of each month, triggered by a scenario. For this, I’ve created a custom SQL trigger that checks if the essential tables are updated by returning 0 or 1. However, this SQL code takes around 15 minutes to execute. To optimize this process, I’d like the trigger…
-
Performance issue using 'Group by' and 'Join' with CSV file
Hi, I have migrated from SAS to Dataiku, but having some performance issue. is there a way to solve the performance issue, using Dataiku for data 'group by' and 'join' recipes with CSV file?
-
Collating evaluation stores.
Hi I have 6 different models and for those models I have 6 different evaluation stores, is there a way to collate all these 6 evaluation stores and create a single collated evaluation store? Operating system used: Windows Operating system used: Windows
-
I need to find string in the text and truncate text untill this string
Operating system used: windows Operating system used: windows
-
dataiku API designer support upload files?
https://community.dataiku.com/discussion/12910/how-to-upload-a-file-through-api-designer is same question like this one, in 2021, You are correct that API services will only accept the query parameters as JSON objects, so uploading a file using multipart/form-data is not possible at the moment. I not sure now is 2025 any…
-
How to install TEXT PREPARATION PLUGIN
Dear Community Friends, Can anybody please help me in solving problem " How to install Text Preparation Plugin step by step ?" including installing dependency modules ,Profibus compiler, Regex Module in Python environment. I am using Windows10 with Dataiku 13 Free Edition . I am repeatedly failed in installing TP. when…
-
While Using Dataiku
Dear Community Friends, Can anybody please help me in solving problem " How to install Text Preparation Plugin step by step ?" including installing dependency modules ,Profibus compiler, Regex Module in Python environment. I am using Windows10 with Dataiku 13 Free Edition . I am repeatedly failed in installing TP. when…
-
Initial install - Connecting this DSS instance to a AKS cluster
Failed to start cluster : <class 'azure.core.exceptions.HttpResponseError'> : (AuthorizationFailed) The client 'xxxx-xxx-xxx-xx' with object id 'xxxx-xxx-xxx-xx' does not have authorization to perform action 'Microsoft.ContainerService/managedClusters/read' over scope…
-
How to Visualize Anomalies in Dataiku Similar to This Chart?
Hi everyone, I'm trying to create a chart in Dataiku to visualize anomalies in my dataset, similar to the example below: The chart highlights three categories: Good Anomaly (blue dots) Expected values (gray dots) Bad Anomaly (red dots) I have the following columns in my dataset: Date/Time (for the x-axis) Sales (for the…
-
delete
n/a
-
List all connection calls
Hi guys, Is there any way to list all calls that a dataiku connections. For exemple, i have a connection pg-teste-op and i need to know many times it was used, that is, how many datasets use this connection and the highest possble level of detail, such as wich job used the connection.
-
[Samsung Fire & Marine] Action is needed to prevent logins from sessions logged in from other IPs
If a session logged in from an IP address called A is tampered with by a user logged in from an IP address called B through the developer tool in IE Edge, the user information will be changed. This needs to be improved as it risks allowing regular users to escalate their privileges to administrator status and manipulate…
-
[Samsung Fire & Marine] Need to improve the performance of Join and Group recipes
Samsung Fire & Marine Insurance has been using a statistical analysis tool called SAS for the past several years. This time I'm trying to replace SAS with Dataiku, but there is a major obstacle and that is the performance issue of Join and Group recipes. When performing tests, when performing a join on about 10 million…
-
Ollama on DSS
Pardon my ignorance on the subject but is there a way to connect to a model like llama2 or mistral running on Ollama locally or another k8s cluster to DSS Operating system used: Suse 15
-
How to Automate Clustering with Anomaly Detection for Each Partition in Dataiku?
Hello Dataiku Community, I’m working on a project where I’ve partitioned my dataset by category and year. For example, my partitions look like this: Category A | 2021 Category A | 2022 Category A | 2023 Category A | 2024 Category B | 2021 Category B | 2022 Category B | 2023 Category B | 2024 Category C | 2021 Category C |…
-
How to Handle Missing Data for Seasonal Analysis in Dataiku?
Hi Dataiku Community, I’m working on a dataset containing daily commodity prices over multiple years (2021-2024). However, there are significant gaps in the data, which are affecting my ability to analyze trends. Here are the details: Daily Data The dataset records daily prices for various commodities, but many days are…
-
Running the script in analysis using the API
I have multiple dataiku models deployed in the flow. I want to refresh the model periodically using the API. The currently deployed model has custom features defined inside the analysis section of the lab. When refreshing the model, I want to copy these steps in the deployed model to the current model trial, create the…
-
How can I change the default location where a .conf file is created to any custom location?
Hello Community, I am using the great expectations package in a dss project for my Data Quality checks. I have already installed it in my code env and using it in a python script for the time being. Even though, the package properly runs in a python notebook when I save it back to recipe I get the following error: " Job…
-
Joining the files in 'Join With' recipes
Dear All, Is there any possibility to join more than 2 files at time? Or the Application behaviour itself, work like joining 1st & 2st file first, post the result; need to join with 3rd file. Even the venn diagram in the Join option show only 2 circles not 3 circles. Operating system used: DATAIKU
-
Dropdown list of Presets
Hello All, I want to create a plugin connector. The plugin will have two types of presets for simplicity will be A and B. the connector.json I implemented is this: "params": [ { "name": "preset_type", "label": "Preset Type", "type": "SELECT", "selectChoices": [ { "value": "A", "label": "A" }, { "value": "B", "label": "B" }…
-
Automatic refresh or reload of source file from local /shared drive path in pc
I need to use DataIKU as ETL tool. We have uploaded our source data of excel format by using 'Upload your files' option. The issue is whenever the file gets replaced in the root path of our local/shared drive folder in pc, dataiku does not pick the updated file automatically. If there is any way to resolve this or another…
-
Generate encrypted passwords manually
I'm trying to set up automatic configuration in my dataiku instance. I can do that through overwriting the files in "dss_dir/config". One issue I'm currently having is how to generate encrypted passwords, for example in the connections file we have: This is "password"…
-
Dataiku password security
Do Dataiku use AES-256 encryption to store the 3rd party passwords ? also is there any way to configure these algorithms Also how frequently the encryption keys are rotated ?
-
Extend the "Rebuild Code Studio templates" option to non-admins when updating a code environment
I was pleasantly surprised to discover the "Rebuild Code Studio templates" option in the "Containerized Execution" settings of a code environment. This feature enables the rebuilding of Code Studio templates that rely on a given code environment, effectively killing two birds with one stone. However, after investigating…
-
lanchpaud
Hello, I don't include in launchpaud, please help me
-
Dataiku users from Romania
Are there any other community members from Romania? PS: > also started this thread as a log for tracing personal progress. >Day one-13.01.2025 - Installed application x Data Preparation Quick Start 8 of 8 lessons completed (100%)
-
delete vs drop
Hi, currently I am working on a project which I have to collect data but it should be new and have the same schema with previous one but without keeping the past data how can I do this? Firstly, I thought that delete option will work for me but at the end, it didn't do what I am looking for; Thanks in advance Operating…
-
Got the error "API Key not provided" when call to python API endpoint
I have created a python function API endpoint in API designer, when I run the test query, it works well. After that, I created an API key for this API service, but when I call the public API endpoint via Postman (i found the guide here: https://doc.dataiku.com/dss/latest/apinode/first-service-apideployer.html) : URL:…
-
Can we use multiple data sources/tables to create knowledge bank for a RAG model?
-
My dashboard is not working it is showing the 404 error
-
evaulate LLM recipe
I want to evaluate LLM like the picture above, but if you look at the tutorial, there is a recipe called evaluate LLM, but I can't see the recipe. How do I evaluate LLM?
-
Updating models in scenarios
Hello, I set up a model with DSS machine learning modules. In my scenario, I've made a special step to run the model training queries. Will this step re-train the model and modify the model coefficients and metrics of the model or not?
-
Correcting Typos - Text Preparation Plugin?
Hi everyone, Context: I have data from a survey. One question in the survey is multiple choice and answers predefined, but there is no data validation built into the survey. As a result, I have typos in the data. For example, a column "Genre" can include "Rcok, Clasisc, Jaz". Question: Is there are smart/quick way to…
-
Error when using Embed recipe
Hi Team, When I am trying to embed a dataset in Dataiku with textual data, I encounter this error: Error in Python process: <class 'Exception'>: At least one embedding request failed: HTTP code: 429, response: { "statusCode": 429, "message": "Token limit is exceeded. Try again in 48 seconds." } Could you guys please help…
-
K-Modes supported
Hi, I noticed that Dataiku supports K-means clustering but couldn't find support for k-modes. Am I missing some documentation. If not, are there any plans to support k-modes clustering? thx Operating system used: Windows 10
-
Questions on quick modeling prediction
I have questions on the quick modeling part of dataiku. Now I am completing an assignment, but I find that the column of data in my label data used to calculate the cost does not appear in the unlabeled data. This problem caused me to be unable to predict unlabeled data with the model I trained. I would like to ask how to…
-
What am I doing wrong with a simple date sort in the prepare recipe?
I have dates that are not parsed. It starts with a group of dates from 2015-01-01 to 2024-12-01 and keeps repeating over and over for each categorical variable. I try to sort the date as is which is as a string and nothing happens. I parse the date (and get a bunch of minute/second junk at the end (ill never want that ever…
-
Error on Extra Tree model in DSS 12
Hi everyone! I'm getting the following error trying to run the XT model. How can I fix it?
-
SAP HANA with Dataiku
Hi everyone, How can I connect to SAP HANA in Dataiku?
-
Handling Empty or Missing Datasets Dynamically in Dataiku Python Recipes
Hi everyone, I’m working on a dynamic process in Dataiku where I pull data from Athena using SQL, then do athena unload, and use the resulting dataset as an input for a pyspark recipe. The challenge is that for certain scenarios (e.g., specific therapeutic areas), the Athena query results in an empty dataset. So I want it…
-
illegalargumentexception: Example has 2 inputs but prompt has 1 input
Hi I'm new to dataiku. I'm using dataiku 13.1. I have a dataset with Area name & I have to find the country in which the area belongs to. I have to do it using AI prompt. My dataset is around 2100 records. I want one country value for each region. So in my Prompt & LLM settings I'm providing prompt as "Find the country in…
-
How to do auto update of the "Variable display" tile in the Application designer
I am using the project variables to be displayed in the "Variable display" tile, Even after updating the project variables, "variable display" tile is not displaying the updated values of those project variables. I need to explicitly refresh the browser to see the "Variable display" tile to display the latest values. How…
-
Improve scenario history : Make them usefull to compare step change
As we know, on DSS we have a Project version control as a builtin "Git-based" version control and we have a kind of lite version of that for any recipe and editable object as known as “History”. Which inside we can check each commit and compare them easily. This seems working well for any kind of recipe with the various…
-
How to get metadata from managed folder of dataiku?
I am trying to get the last modification from the metadata of managed folder of dataiku but it is throwing below error Exception: None: b"Failed to convert value of type 'java.lang.String' to required type 'boolean'; nested exception is java.lang.IllegalArgumentException: Invalid boolean value [/GSX Daily Pos.xlsx], caused…
-
Issue with Rendering Images in HTML from Temporary Folder path in Dataiku
Hi everyone, I am facing issue while trying to render images in a HTML file that i have generated from a word document using mammoth library in Python in dataiku. Here's what i have done so far: a) Extracted images from the word document and saved them in to a managed folder in dataiku named "images" as i didn't want to go…
-
Automate project deployments to Prod Instance
Hello all, I’m exploring ways to optimize and automate processes within Dataiku and have a few questions: Is there a way to automate the deployment of projects to Ops (operational environments) within Dataiku? If so, what are the best practices or tools/plugins to achieve this? What are some recommended approaches for…
-
Design node and API node is powering since half an hour.
I am trying to opening a Dataiku Launchpad. After turning on The design node and API node is powering since 45 minutes now. Is that a normal processing time or there is some error? What is the solution to it? Operating system used: WIndows
-
Somebody can help me in this regard
Hi Dataiku Community , Please provide a solution for this problem which I am facing for a long time Operating system used: Windows 10
-
How do I import data from Dataiku to SAP Hana
Hi I'm new to Dataiku. I'm using Dataiku version 13.1. SAP Hana integration is done with my dataiku. I'm using SAP Hana dataset for my ML use case. I need to import my ML output table to SAP Hana. Can you please provide me steps to import data from dataiku to SAP Hana Operating system used: Windows Operating system used:…
-
Can we connect to SAP Hana using JDBC driver using Dataiku?
-
Writing to partitioned dataset using the writer
I'm trying to overwrite a table using data from another table (with the same schema). I keep running into the issue that both datasets are partitioned and the writer does not like that (same case with the copy_to function). Here is what I'm trying to do: ERROR:dataiku.core.dataset_write:Exception caught while…
-
Windows recipe
Hello, I am currently working on a window recipe (to add a row number), and I would like this row number to be in the first column without using a Prepare recipe exclusively for that. Do you think it is possible ? Thank you !
-
How often does elapsed time on a job update?
On the list of for currently running jobs in dataiku, how frequently does the elapsed time on a running job update to reflect that more time has passed? Is it less frequent for long jobs?
-
How to accumulate numbers from largest to smallest
As sample data shows, the data has customer ID column and revenue by customer column, and I would like to accumulate the revenue from largest revenue to smallest revenue to identify which customers are in top80% market and which customers are in bottom 20%. How can I calculate like orange-colored column in DSS?
-
Configurable Timezone Display for Date Columns (Beyond UTC-only)
Current Situation Dataiku DSS has specific behaviors when handling time columns: When it recognizes time-related columns (e.g., date, timestamp_tz, or timestamp_ntz), it displays them as Date columns, rendering them in timestamp format (with both date and time components). A significant limitation is that Date columns…
-
PYSPARK_PYTHON environment variable issue in PySpark
Hi I am facing below issue for PySpark recipe. Exception: Python in worker has different version 2.7 than that in driver 3.6, PySpark cannot run with different minor versions.Please check environment variables PYSPARK_PYTHON and PYSPARK_DRIVER_PYTHON are correctly set. I have set the environment variables using…