-
Issues with pandas version
Hello, I am evaluating DSS. I am not expert of it, I am trying to load a project that was previously created by some colleagues in a previous version, I think it was 4.0 I am able to import the project (which contains hadoop and spark steps). The problem is when I try do build all the flow. I am receiving this error:…
-
Cannot run program "conda" error in DSS code environment settings.
Hi, I have both anaconda and python installed in GCP instance where DSS is also installed. Conda was working fine initially and i was able to update and install package using both Conda and pip. However after 3 weeks , i am getting this error " Cannot run program "conda": error=2, No such file or directory, caused by:…
-
Newly occurring Jupyter Notebook error
I have several python jupyter notebooks across a few projects on our DSS install. I recently upgraded our server to 4.3.0. When I try to open a notebook I get the message: Action denied If I open the developer console I see that I'm getting a 403 when the system tries to send a request to /jupyter/api/contents/. I've…
-
Delete partitions in DSSManagedFolder
Hi, having an object handler of type dataikuapi.dss.managedfolder.DSSManagedFolder how can I remove all partitions except the last N ones? The dataiku.core.managed_folder.Folder has nice methods list_partitions() and clear_partition(), but this one does not provide. Should I get the definition of partition scheme, use…
-
Capturing Python logging output in a file
Looking for a way to capture Python logging output from a recipe in a file. I explored directing output to a file in a managed folder but writing to a managed folder requires use of the API. Any other ideas?
-
Error when executing SQLExecutor
I got an error when executing SQLexecutor function in Dataiku IPyhton. This is the command that I used: import dataiku import pandas as pd, numpy as np from dataiku import pandasutils as pdu from dataiku.core.sql import SQLExecutor2 from dataiku.core.sql import SQLExecutor dtk =…
-
Is it possible to make calls to OpenRefine Server from DSS?
Hi, Is it possible to make calls to DSS's Open Refine Server. We want to make connection to Open Refine server that DSS is already using, in "Python Recipe" instead of "Prepare Recipe" Please let us know how we can do it. Waiting for your kind response. Regards, Samriddhi
-
delete dataset without clearing the metrics values
Hi, I am using the python public API to clear the dataset. This deletes the data, but also the history of the metric values. Is it possible to delete just the data? And one more related question, when the dataset is linked to the folder, the clear() method does not remove the data. Do I have to look up for the folder and…
-
rename dataset by python internal API
Hello everyone! I'm working on a Python Notebook to copy paste a Project. I have already change the path of my imported Datasets like explained here: https://www.dataiku.com/learn/guide/tips/duplicate-project.html. However I can't rename the dataset like I renamed the 'hiveTableName'. Any help please? :) #Foreach datasets…
-
Create sync recipe with python code
Hi, how can I create a impala sync recipe with the Public API? I have the source managed dataset stored as parquet and I would like to create a code recipe with sql like "select count(*) from mytable" into a new parquet managed dataset. So far I have been using this, but this method assumes that the output dataset is…