Using Dataiku

Sort by:
21 - 30 of 291
  • Hi, I am using the local deployer to deploy project on automation node. The first deployment goes well, on the other hand when I try to do an update and deploy a second version i get this error messag…
    Answered
    Started by Mahran
    Most recent by Grixis
    0
    6
    Grixis
    Last answer by Grixis

    I think you should go back to your deployer node and go to your project's bundle deployment interface and go to the update logs and look at the traceback in more detail because you'll have more information. I don't see how I can help you without more information. :(

    sometimes you have this kind of error because the bundle version doesnt match with the current project deployed.

    Grixis
    Last answer by Grixis

    I think you should go back to your deployer node and go to your project's bundle deployment interface and go to the update logs and look at the traceback in more detail because you'll have more information. I don't see how I can help you without more information. :(

    sometimes you have this kind of error because the bundle version doesnt match with the current project deployed.

  • Dataiku Version: 13.3.1 I have several LLMs that I have in my DSS cache from the HuggingFace Connection. I can leverage these models using the prompt recipe. However I am struggling to use them in cus…
    Question
    Started by areisdorf
    0
  • Hi, can someone help me please. Given a folder input and a folder output I want to link them with a shell script so that the shell script can read a test.txt file from input folder and write the outpu…
    Answered
    Started by Klajdi
    Most recent by Turribeach
    0
    6
    Last answer by
    Turribeach
    Last answer by Turribeach

    Sorry but this is a horrible solution. I am sure there are much better alternatives. Why can't you use a standard Sync recipe to copy the files between folders? Why are you not allowed to use Python? Dataiku without access to Python is pretty much useless. This doesn't sound right, there are aspects that you are not really clarifying. Can you explain the requirement in detail?

  • How do I identify a record where the value in a specific field is NULL, in a Recipe filter condition?
    Answered ✓
    Started by Dbase3tate
    Most recent by Dbase3tate
    0
    4
    Solution by
    Dbase3tate
    Solution by Dbase3tate

    Thanks, I think that worked. I need to research but it looks a lot better.

  • A new scenario step was added in a recent DSS version which is to execute a Python unit test. I'd like to start using this. However, the documentation is pretty brief: "this step executes one or more …
    Answered ✓
    Started by Marlan
    Most recent by Marlan
    1
    3
    Solution by
    Marlan
    Solution by Marlan

    • I found a tutorial that shows how to use the Python Unit test functionality in the Knowledge Base:
  • Hello, I'm currently using Dataiku and SQLExecutor2 to run queries on my Oracle database hosted on AWS RDS, port 2484. When I execute the same query from a Docker container on AWS, the query takes abo…
    Answered
    Started by pPGrillo
    Most recent by Turribeach
    0
    1
    Last answer by
    Turribeach
    Last answer by Turribeach

    We don’t have all the facts to compare. You say you are running the queries in Dataiku, what specific Oracle driver version are you using in Dataiku? What Oracle database version do you connect to? Where is your DSD node located? Where are your docker containers located? What exactly do you mean by “I execute the same query from a Docker container on AWS”? You need to describe the whole technology stack there including all the driver and software stack versions used.

    Finally how do you do go from a query that takes 1 second to run to taking 2 days? Are you running SQL queries in a for loop? That’s a really bad pattern. Explain exactly what is your requirement and how you are trying to achieve it.

  • I have a scenario that builds and exports a Dashboard and survived several Dataiku version upgrades until now. When the scenario runs I have a Reporter that exports the Dashboard into a PDFs and sends…
    Answered ✓
    Started by me2
    Most recent by me2
    0
    2
    Solution by
    Turribeach
    Solution by Turribeach

    This is likely to be a problem in your email addresses in your reporter. Do you have any special characters, spaces or anything weird in them?

  • Hi all, dataiku novice here. I have trained models in python notebooks with mlflow, then I deployed a model to the flow(from experiment tracking) in order to score & evaluate it. I have 50 features, a…
    Answered
    Started by esmvy
    Most recent by tgb417
    0
    3
    Last answer by
    tgb417
    Last answer by tgb417

    @esmvy ,

    I don't have a good answer for you at this time. And I don't have a bunch of time to followup. Several thoughts.

    1. I've seen problems like this when using visual models. I'd add a column to datasets prior to the model inference. Then I'd have to regenerate the model to include the new column and re-publish the model. Then I could run the model on the new dataset.
    2. Based on the Python you are showing, it appears that you are not using the visual model building approach. I've not build models in Dataiku that way, so, I'm not clear what you might be doing. Others may be able to understand exactly what you are doing from this bit of code. However, I am not.
    3. If I were you, I would likely open up a support ticket. The Technical support team at Dataiku are often very good to excellent. They can get some diagnostics and more accurately identify the root cause.

    —Tom

  • I'm creating a python function endpoint with this script: And I don't know how to deal with this error: Dev server deployment FAILED Failed to initiate function server : <class 'Exception'> : Default …
    Answered
    Started by Eduardo
    Most recent by Velichka
    0
    2
    Last answer by
    Velichka
    Last answer by Velichka

    Hello all,

    i am still looking for a solution to my problem. I have the following Jupyter Notebook:

    import dataiku
    import pickle
    import pandas as pd
    from sklearn.feature_extraction.text import TfidfVectorizer
    from sklearn.metrics.pairwise import cosine_similarity
    import dataikuapi

    def load_data(folder_name="Recommender"):
    managed_folder = dataiku.Folder(folder_name)


    with managed_folder.get_download_stream("cosine_similarity.pkl") as stream:
    cosine_sim = pickle.load(stream)


    with managed_folder.get_download_stream("tfidf_vectorizer.pkl") as stream:
    vectorizer = pickle.load(stream)

    with managed_folder.get_download_stream("tfidf_matrix.pkl") as stream:
    X_tfidf = pickle.load(stream)


    sachnummer = dataiku.Dataset("LTM_prep")
    df = sachnummer.get_dataframe()
    df.drop(['lieferant_name', 'lieferant_ort', 'LIEFERANT_NAME_ORT', 'LT_GEBINDE_NUMMER', 'MDI'], axis=1, inplace=True)

    return cosine_sim, vectorizer, X_tfidf, df

    def recommend_filtered1(input_bennenung, vectorizer, X_tfidf, df, top_n=10):
    try:
    if not input_bennenung:
    return {"error": "Die Eingabe-Benennung darf nicht leer sein."}

    input_bennenung = input_bennenung.upper()
    input_vector = vectorizer.transform([input_bennenung])
    similarities = cosine_similarity(input_vector, X_tfidf).flatten()

    top_indices = similarities.argsort()[-top_n:][::-1]
    recommendations = [
    {"test": df.iloc[idx]['test'],
    "test2": df.iloc[idx]['test2'],
    "SIMILARITY_SCORE": round(similarities[idx], 2)}
    for idx in top_indices if similarities[idx] > 0
    ]

    return recommendations if recommendations else {"message": "Keine ähnlichen Benennungen gefunden."}
    except Exception as e:
    return {"error": f"Fehler: {str(e)}"}

    def recommend_from_input(input_bennenung):

    folder_name = "Recommender"

    if not input_bennenung:
    return {"error": "Fehlender Parameter 'input_bennenung'"}

    try:
    # Lade alle benötigten Objekte
    cosine_sim, vectorizer, X_tfidf, df = load_data(folder_name)

    # Empfehlung berechnen
    return recommend_filtered1(input_bennenung, vectorizer, X_tfidf, df)
    except Exception as e:
    return {"error": f"Fehler beim Laden der Daten oder der Empfehlung: {str(e)}"}

    and want to call the method recommend_from_input from it. I am in the API Designer. I have a managed folder called "Recommender," which I can also see in the Flow. The structure im Folder is

    Unter Folder Settings i see type is Amazon S3 and i have a setted connection and see also the path in bucket. So when i call def recommend_from_input(input_bennenung): return input_bennenung in the api designer code section with the test query

    {
    "input_bennenung": "Stern"
    }

    there are no errors and i get "Stern back". So now i just pasted my notebook code in the api designer code Section and when i run it there is a error:

    Result: 

    {"error":"Fehler beim Laden der Daten oder der Empfehlung: Default project key is not specified (no DKU_CURRENT_PROJECT_KEY in env)"}

    In the logs are no errors only info and debug.

    I would appreciate any help. I have already read the documentation on Exposing Python Functions, but I still don't know where my mistake is.

21 - 30 of 2913