-
Add Seldon to deployment options
One of the deployment options in our company is Seldon (Seldon, MLOps for the Enterprise.). It would be great if Dataiku had the option to deploy directly to Seldon, the way deployment to K8, AWS, Databricks or Azure is now possible. Seldon in general deploys MLflow artefacts.
-
Put stuff in the API logging without sending in the response
We often run into situations where we'd like to log stuff from our internal API workings - like intermediary results for checking - without having to send these out in the response. It would be wonderful if there was an option to send things to the API log without it having to be part of either the request or the response.
-
HTTP status codes returned by DSS API node/service
Two questions: 1. Will an exception, raised during execution of API service code, always result in a response with a 500 HTTP status code? 2. Aside from that, errors can occur outside of API service code execution (e.g. service is unavailable). Is there an overview available of which situations map to which HTTP status…
-
Default project key is not specified (no DKU_CURRENT_PROJECT_KEY in env)
I'm creating a python function endpoint with this script: And I don't know how to deal with this error: Dev server deployment FAILED Failed to initiate function server : <class 'Exception'> : Default project key is not specified (no DKU_CURRENT_PROJECT_KEY in env)
-
What is the difference between the dataiku and the dataikuapi packages?
From what I could see, both seem to have the same functionality when I used them outside the Dataiku web interface. I see documentation for both the packages but they seem to be conflicting. For example, in this page section, in the illustrative code, we are importing dataiku, but just below it, it is written that a…
-
Dataiku Spark Remote Connection
Hello, I am using Dataiku 12.5.2 and currently running Spark 2.4. Dataiku is installed on a server named A, while Spark is installed on a server named B, configured as a standalone Spark installation without Hadoop. Both server A and server B are capable of TCP communication and allow SSH access. How can I use the Spark on…
-
remapping connections for API services
Goodday! In the API Designer, we can define connections to use with SQL Query Endpoints. How do we remap these connections based on deployments to different API nodes? (ie. use different connection for deployments to a production API node vs. deployments to an acceptance API node) I don't see any option in the deployer UI…
-
LLM
Hi , Is there a way to use LLM connections and models as a API node? Operating system used: Windows
-
Failed: Failed to run function : : No DSS URL or API key found from any location
Hi i am getting an error while doing the TESTING QUERIES on the API designer for python function endpoint. Failed: Failed to run function : class 'Exception' : No DSS URL or API key found from any location
-
Use file from code env resources directory with API Test queries
I've built a python API endpoint that uses a certain code environment. In the resources folder of that code environment is a file. If I run the API code in a notebook using that code env, I'm able to access that file. But when I run Test queries in the API Designer, the file cannot be found. The code used is the same in…