Using Dataiku

Sort by:
61 - 70 of 93
  • Hello all, is it possible to round off numbers using the Formula language. Example 0.6 -> 0 I have only found the classic rounding and the rounding to even or odd numbers. Thanks for your help.
    Answered ✓
    Started by Max_M
    Most recent by Max_M
    0
    2
    Sergey
    Solution by Sergey

    Hi @Max_M

    Do you need to round all the decimals to integers using the floor function? This is available by default. So this data:

    first,second
    0.3,0.6
    1.2,2.3
    3.6,3.8

    will be transformed this way:

    Screenshot 2022-03-29 at 19.45.31.png

    Sergey
    Solution by Sergey

    Hi @Max_M

    Do you need to round all the decimals to integers using the floor function? This is available by default. So this data:

    first,second
    0.3,0.6
    1.2,2.3
    3.6,3.8

    will be transformed this way:

    Screenshot 2022-03-29 at 19.45.31.png

  • Hi Team, We have a case where we need to update all our recipe codes(SQL/Python) to modify the existing database name and schema name to a new one. Unfortunately, the existing code is a pain as we hav…
    Question
    Started by sridarvenkatesa
    Most recent by Marlan
    0
    1
    Marlan
    Last answer by Marlan

    Hi @sridarvenkatesa
    ,

    Your thought is to export the project to a zip file, unzip it, make the changes (using a search and replace tool presumably), and then zip and reimport it?

    I don't know for sure but I'd think that could work.

    You could also use the Python API (dataikuapi) to make the changes in place. Let me know if this is of interest as I have some code you could modify to do this. We are using this approach to partially convert SQL recipes in our projects from Netezza SQL to Snowflake SQL.

    Marlan

    Marlan
    Last answer by Marlan

    Hi @sridarvenkatesa
    ,

    Your thought is to export the project to a zip file, unzip it, make the changes (using a search and replace tool presumably), and then zip and reimport it?

    I don't know for sure but I'd think that could work.

    You could also use the Python API (dataikuapi) to make the changes in place. Let me know if this is of interest as I have some code you could modify to do this. We are using this approach to partially convert SQL recipes in our projects from Netezza SQL to Snowflake SQL.

    Marlan

  • AttributeError: module 'dataikuapi' has no attribute 'DSSClient' using the above code im trying to simply test connecting to DSS instance from python. Python is on my local PC and DSS is on AWS. Insta…
    Answered ✓
    Started by Von1974
    Most recent by CoreyS
    0
    2
    Solution by
    Von1974
    Solution by Von1974

    found the problem. Installed libs as root and logged on as another user.

  • I am trying to read an Oracle table with partitioning. I have specified a date column as a partition but I am getting the following error: Failed to read data from table Failed to read data from table…
    Question
    Started by PARTEEK
    Most recent by fchataigner2
    0
    1
    Last answer by
    fchataigner2
    Last answer by fchataigner2

    Hi,

    DSS apparently can't use the column type of that date for partitioning. The simplest is to make a view out of the table and use the view instead of the original table in your DSS dataset. Something like

    create view xxxxx as select cast(date_col as NVARCHAR(50)), ... from the_table

    (or some TO_CHAR(...) with an appropriate format)

  • Hi there, In the model design tab, if I select a specific Feature Reduction technique, then how can I see the final list of features that were used to train the model? I can see all the input features…
    Question
    Started by asadmasad
    0
  • Is it possible to use a directory as an input for a data source? I need to pull in all files within a directory. They all have the same schema.
    Question
    Started by MClark05
    Most recent by Sergey
    0
    1
    Last answer by
    Sergey
    Last answer by Sergey

    Hi @MClark05

    Yes, you can create an FS connection pointing to this directory and create a managed folder based on it. When done, use python APIs listed below the mentioned doc page to read the data from it.

  • I know DSS only supports XLSX, not XLSB! Is there any alternative ways to load XLSB (binary) format? Also wondering why DSS is not supporting XLSB Format Cheers! Naresh
    Question
    Started by tinaresh
    Most recent by CoreyS
    0
    2
    Last answer by
    CoreyS
    Last answer by CoreyS

    Hi @tinaresh
    please feel free to utilize the Product Ideas board. The Product Ideas board is here to let you share and exchange your ideas on how to improve Dataiku. Here are some resources to help get you started: Suggest an idea

    I hope this helps!

  • Hello, I have a project where I created a Managed folder where i succeeded to load a text file and binary file in it. To reach the content of my folder , i use: REPO = dataiku.Folder("quNauGux") REPO …
    Question
    Started by shoareau
    Most recent by Keiji
    0
    1
    Last answer by
    Keiji
    Last answer by Keiji

    Hello @shoareau
    ,

    Thank you for posting the question on Community.

    > How can i get the absolute path of the file inside DDS ?

    If your managed folder is stored on the local filesystem of the DSS host, you can use the Folder.file_path(filename) function to extract the filesystem path for a given file within a folder. Here is sample code.

    import dataiku
    
    FOLDER_ID = 'VdrA1ZMC'
    
    REPO = dataiku.Folder(FOLDER_ID)
    list_files = REPO .list_paths_in_partition()
    REPO.file_path(list_files[0])

    Screen Shot 2022-03-27 at 12.06.33.png

    Please note that this Folder.file_path(filename) function is available only if your managed folder is stored on the local filesystem of the DSS host. If your managed folder is stored on other places such as HDFS and S3, this function cannot be used and you will need to call the Folder.get_download_stream(path) function to retrieve the file contents from the folder as follows.

    import dataiku
    
    FOLDER_ID = 'RoQHOCas'
    
    REPO = dataiku.Folder(FOLDER_ID)
    list_files = REPO .list_paths_in_partition()
    with REPO.get_download_stream(list_files[0]) as stream:
        data = stream.read()
        # You can handle / manipulate the file contents (data) here.

    I hope this would help. Please let us know if you have any further questions.

    Sincerely,
    Keiji, Dataiku Technical Support

  • Hi, I am doing a filter recipe in order to keep distinct values and drop the duplicates, from two different columns using a formula. How can I do that on DSS? Or what is the equivalent to "ne" in dss?…
    Question
    Started by pinpin_3
    Most recent by Manuel
    0
    1
    Last answer by
    Manuel
    Last answer by Manuel

    Hi,

    You can use a Group recipe with those two columns as keys.

    If there are more columns, you need to decide what to do with them and use the aggregation section of the recipe for that.

    I hope this helps.

    Best regards

  • Hi Guys, Can you pls. help me with if i have 35 scenario and all 35 scenario dependent on another various individual respective 35 individual scenario. How can i set this up?
    Question
    Started by dave
    Most recent by EliasH
    0
    5
    Last answer by
    EliasH
    Last answer by EliasH

    Hi @dave
    ,

    Ahh thank you for the example So I'm assuming when you say "Business" you're referring to a project? Sounds like what you need to use is a "Trigger after Scenario" trigger.

    In your example, in the settings of Business Y Scenario A1, create a "Trigger after Scenario" trigger where the "Scenario" to trigger after each run is Business X Scenario A.

    Scenario A1 can run when Scenario A either finishes, runs successful, produces warnings, fails, or is aborted.

    Screen Shot 2022-03-25 at 3.30.01 PM.png

    So just to clarify a bit more, in the "Scenario" field of Scenario A1 in Business Y, you would select "Scenario A (BUSINESSX)" and select "runs successfully" in the "when the other scenario..." field.

61 - 70 of 937