-
Calculate a single metric/check via the Dataiku API
Hi, It is currently not possible to calculate a single metric or check via the Dataiku API while this is possible via the GUI. The following APIs exists: dataset.compute_metrics() dataset.run_checks() but they will calculate all enabled metrics/checks which may take a lot of time. So this idea is to provide an API to allow…
-
Edit default metrics and checks as a project-wide setting
When creating a new dataset, I practically always edit the default metrics and checks to run row counts after build. Ideally, I could define this from the project settings so that every new dataset created automatically has my desired metrics and checks configured. Of course, this doesn't apply to column-specific values,…
-
Add Seldon to deployment options
One of the deployment options in our company is Seldon (Seldon, MLOps for the Enterprise.). It would be great if Dataiku had the option to deploy directly to Seldon, the way deployment to K8, AWS, Databricks or Azure is now possible. Seldon in general deploys MLflow artefacts.
-
Add support for Snowflake key-pair authentication
Currently the dataiku only support "User/Password" and "OAuth" for Snowflake connection. Snowflake has Key-pair authentication https://docs.snowflake.com/en/user-guide/key-pair-auth, which I would like to use for my service accounts.
-
Field grouping
Hello, A lot of time, when you have a dataset, you want to know if there is a group of fields that works together. That can help to normalize (like de-joining) your data model for dataviz, performance issue or simplify your analysis. Exemple <style></style> order_id item_id label model_id length color amount 1 1 A 10 15…
-
Add/Delete rows button for VisualEdit Plugin
The visual edit plugin is really good and does what we need, however I am having to code a dash webapp that uses the visual edit class and add my own functions for adding and deleting rows in the editable dataset, it would be great if you can configure the visual edit plugin to do this instead of having to custom code it.
-
Present Connection Credentials in sorted order
Connection Credentials are currently presented to the user in what seems to be a random order. That combined with the fact that we have about 70 connections defined makes it pretty difficult to find a particular connection. It would be great if the connections were sorted as they are in most other areas of DSS. For…
-
Full Screen Streamlit Apps (stand-alone view through Dataiku)
The title says it all. Publishing your Streamlit Webapp via Dataiku feels more like a clunky embedded experience than a bespoke application. What I'm looking for here is a native experience through Dataiku's platform. I know that Dataiku offers users the ability to publish their web apps publicly (full screen / native…
-
Databricks as pyspark engine
Currently, Dataiku only supports Databricks using Databricks Connect (in a python recipe), but does not support pyspark recipes or selecting spark engine for visual recipes. As a result, we are not able to develop visual recipes with spark, and either need to use Databricks connect in python recipes, or develop directly in…
-
export dataiku plots directly as svg
It would be great if users could easily download all plots/charts created in dataiku directly as an svg (including model performance plots). For users who code, it would be even better if you could open an ipynb notebook with the code for generating the plot so the user could customize it and then export in a…