-
Integration with Microsoft Fabric and its OneLake
Hi, Couldn't find anything on an integration with Microsoft Fabric through OneLake yet in the Dataiku docs/release notes. Is this coming soon? As I read the Microsoft docs I understand we can't connect directly via ADLS, only via APIs or SDKs. Thanks in advance, Jonathyan Operating system used: RHEL 8
-
Remapping Connection (under Activation Settings) in Bundle Management using Dataiku Python API
I am trying to remap/update the already existing Blob container connection name I have created a Bundle in Design Node and imported & activated in Auto Node. Both the nodes are pointing to separate Azure Blob Containers Unable to figure out if the connection settings have to be updated before the creation of the bundle OR…
-
How to output to / update my snowflake table using Dataiku
I have a snowflake table and I've set up the connection and everything looks good, Dataiku requires me to create a dataset using that snowflake table that I can use as my input / output. The issue is I have that dataset as my output and when I run my flow, I can see my results, but it isn't actually outputting to my…
-
Dataiku Spark Remote Connection
Hello, I am using Dataiku 12.5.2 and currently running Spark 2.4. Dataiku is installed on a server named A, while Spark is installed on a server named B, configured as a standalone Spark installation without Hadoop. Both server A and server B are capable of TCP communication and allow SSH access. How can I use the Spark on…
-
remapping connections for API services
Goodday! In the API Designer, we can define connections to use with SQL Query Endpoints. How do we remap these connections based on deployments to different API nodes? (ie. use different connection for deployments to a production API node vs. deployments to an acceptance API node) I don't see any option in the deployer UI…
-
Connection to locally hosted LLM on Apple Silicon
Is there any way to setup a connection that connects to a locally running LLM? On Apple Silicon it is possible to run huggingface models locally with some different open source solutions, which make use of Metal for the GPU-compute. The solution I have running can be said to function as a server accessible as localhost.…
-
Dynamic rename and addition/removal of columns
I am connecting to QuickBase using a URL path and looking for a way to dynamically add or remove columns from the initial schema as well as rename the columns as "Simplify (and lower case)". I do not own the source data, and cannot influence the naming convention or columns at the source. I am pushing this data to GCP/Big…
-
Is there a list of all the builtin variables like ${projectKey}, ${dssUserLogin}, etc
I noticed the user of some builtin variables in the documentation like: * projectKey jobProjectKey * dssUserLogin * jobProjectKey * jobId * activityId but it's there a list somewhere with all the possibilities, and where they can be used? For example is not clear to me if I can use ${jobProjectKey} in the Administration >…
-
How to unzip files in a remote Managed Folder (as Azure File Share) using python?
I have zip files in a remote Managed Folder (as Azure File Share) and I need to unzip these files using Dataiku (Python) in this same folder to proceed with the flow. However, I was not able to do it following the examples showed in the API documentation (Managed folders — Dataiku DSS 12 documentation) Is it possible to do…
-
AWS Bedrock Connectivity Issue
Hi, I am able to connect to AWS S3 bucket with Access key and Secret key. But When I try to connect AWS bedrock using the s3 connection I am getting error. I have AWS bedrock Full Access permissions. I Have access to two models in AWS. I am trying to connect those model by selecting only them, But even then I am getting…