Installation / DSS Start failure

PGuti Registered Posts: 5 ✭✭✭✭✭
edited July 16 in Using Dataiku


I had a quite old instance of DSS on my mac (I used it last time in late 2017).

I wanted to run it but I got an error when trying to launch it.

So I thought OK I will just download the latest version and upgrade it.

I did this and when I'm trying to start DSS I get the following error:

[Install] ******************************************
* Attempting to rollback
Backing up partially-migrated: dss-version.json -> /Users/pierregutierrez/Library/DataScienceStudio/dss_home/_broken_migration_backup_20211029-103623/dss-version.json

[Install] Restoring backup: _pre_migration_backup_20211029-103623/dss-version.json -> dss-version.json

[Install] Backing up partially-migrated: bin/ -> /Users/pierregutierrez/Library/DataScienceStudio/dss_home/_broken_migration_backup_20211029-103623/
Restoring backup: _pre_migration_backup_20211029-103623/ -> bin/

[Install] Backing up partially-migrated: install.ini -> /Users/pierregutierrez/Library/DataScienceStudio/dss_home/_broken_migration_backup_20211029-103623/install.ini

[Install] Restoring backup: _pre_migration_backup_20211029-103623/install.ini -> install.ini
Backing up partially-migrated: databases -> /Users/pierregutierrez/Library/DataScienceStudio/dss_home/_broken_migration_backup_20211029-103623/databases

[Install] Restoring backup: _pre_migration_backup_20211029-103623/databases -> databases

[Install] Backing up partially-migrated: config -> /Users/pierregutierrez/Library/DataScienceStudio/dss_home/_broken_migration_backup_20211029-103623/config

[Install] Restoring backup: _pre_migration_backup_20211029-103623/config -> config
Restore complete, removing marker file

[Install] Traceback (most recent call last):
  File "/Applications/", line 30, in <module>
  File "/Applications/", line 98, in migrate

[Install]     reraise(*sys.exc_info())
  File "/Applications/", line 67, in migrate
    migration.execute(dip_home, node_type)
  File "/Applications/", line 136, in execute

[Install]     wrapper.execute(diphome)
  File "/Applications/", line 79, in execute
  File "/Applications/", line 132, in execute

[Install]     migrate_json_file(self, fp)
  File "/Applications/", line 95, in migrate_json_file
    new_obj = transformer.transform(cobj, fp)
  File "/Applications/", line 4940, in transform

[Install]     return migrate_ml_spark_params(modeling)
  File "/Applications/", line 4907, in migrate_ml_spark_params

[Install]     spark_params[prop] = obj.get(prop, None)
AttributeError: 'NoneType' object has no attribute 'get'

[install] EOF reached
[install] EOF reached
DSS installer terminated with status 1
DSS installer failed , fatal error

Any idea how to fix this?

Thank you!

Best Answer

  • Huixian
    Huixian Dataiker, Dataiku DSS Core Designer, Dataiku DSS Adv Designer, Registered Posts: 2 Dataiker
    Answer ✓

    Hi @PGuti

    Thank you for the information!

    Can you please check if the params.json file under "/Users/pierregutierrez/Library/DataScienceStudio/dss_home/config/projects/VELIB/analysis/W9NhaPCz/ml/7iRWiMaj/" is null/empty. If so, can you please delete the file or move it outside of the DSS data directory and try again with the migration?

    If the issue persists after that, please attach the full log to the chat you opened, so that we can review and investigate further.

    Thank you!

    Huixian Pan


  • Sergey
    Sergey Dataiker, Dataiku DSS Core Designer, Dataiku DSS & SQL, Dataiku DSS Core Concepts Posts: 365 Dataiker

    Hi @PGuti

    Can you please send us a couple of dozens of lines before this error happened? We need to see the last step the migration was doing before failure.


  • PGuti
    PGuti Registered Posts: 5 ✭✭✭✭✭

    The complete log would show all the project names on a public page which I would prefer avoiding.

    The exact line before was:


    [Install] Migrating /Users/pierregutierrez/Library/DataScienceStudio/dss_home/config/projects/VELIB/analysis/W9NhaPCz/ml/7iRWiMaj/params.json


    The last step written was:


    20 -> 21

    - Migrates spark params in saved models into a common object



    I had a bunch of warning messages:
    [Install] ** WARNING: Unhandled algorithm: KMEANS

    [Install] ** WARNING: Unhandled algorithm: EXTRA_TREES

    [Install] ** WARNING: Unhandled algorithm: NEURAL_NETWORK

    [Install] ** WARNING: Unhandled algorithm: PY_TWO_STEP

    which I'm assume are models which changed and custom models.

    The beginning of the log seems to explicit what is going to be done and is:


    Starting Data Science
    dssHome:/Users/pierregutierrez/Library/DataScienceStudio/dss_home dssBin:/Users/pierregutierrez/Library/DataScienceStudio/dss_home/bin/dss
    Starting Java check
    [Install] [+] Using Java at /Library/Internet Plug-Ins/JavaAppletPlugin.plugin/Contents/Home/bin/java : java version "1.8.0_121"

    [Install] /Library/Internet Plug-Ins/JavaAppletPlugin.plugin/Contents/Home/bin/java

    [Java check] EOF reached
    [Java check] EOF reached
    Java check terminated with status 0
    dss-version exists, doing stop-upgrade-start
    Running DSS task for 'stop'
    [DSS] dss: DSS supervisor is not running

    [dss stop] EOF reached
    [dss stop] EOF reached
    [Install] *********************************************
    * Dataiku DSS installer *

    [Install] [+] /Users/pierregutierrez/Library/DataScienceStudio/dss_home already exists and is a DSS installation

    [Install] [+] Saving installation log to /Users/pierregutierrez/Library/DataScienceStudio/dss_home/run/install.log

    [Install] [+] Using Java at /Library/Internet Plug-Ins/JavaAppletPlugin.plugin/Contents/Home/bin/java : java version "1.8.0_121"

    [Install] [+] Checking required dependencies

    [Install] + Detected OS distribution : osx 11.6

    [Install] + Checking required packages...

    [Install] [+] Migrating Python environment

    [Install] [+] Precompiling Dataiku Python code

    [Install] [+] Precompiling Jupyter Python code

    [Install] [+] Precompiling third-party Python 2.7 code

    [Install] [+] Migrating data directory

    [Install] Select how to migrate from 17 from avail: [8 -> 9, 9 -> 10, 10 -> 11, 11 -> 12, 12 -> 13, 13 -> 14, 14 -> 15, 15 -> 16, 16 -> 17, 17 -> 18, 18 -> 19, 19 -> 20, 20 -> 21, 21 -> 22, 22 -> 23, 23 -> 24, 24 -> 6000, 6000 -> 6020, 6020 -> 6030, 6030 -> 7000, 7000 -> 7020, 7020 -> 8000, 8000 -> 8020, 8999 -> 9000]
    Executing the following migrations

    17 -> 18
    - Clean up the partitioning of SQL datasets
    - Upgrade the y/n resize-to-fit setting of dashboard images to new multi-option setting
    - Cleanup trained models parameters
    - Nest feature generation parameters
    - Drop caches
    - Drop persistent notifications database

    18 -> 19

    19 -> 20
    - Add params to ML recipes
    - Transform comments insight to discussions insights
    - Transform dashboard by changing comments insight to discussions insights
    - Drop caches
    - Drop persistent notifications database


    20 -> 21
    - Migrates spark params in saved models into a common object
    - Migrates spark params in analyses' models into a common object
    - Migrates spark params in analyses' ml tasks into a common object
    - Regroup spark params in recipes
    - Drop caches
    - Drop persistent notifications database

    21 -> 22
    - Give NPS Survey Settings to each user

    22 -> 23
    [Install] Setup metadata for code envs in Jupyter kernels' specs
    - Upgrade structure of dss_schema_info in H2 databases
    - Make CLASS_WEIGHT the default weighting strategy for classification saved models
    - Sets a default Git URL whitelist for clones/pulls
    - Use smartName in refresh charts steps
    - Move ml task param skipExpensiveReports in modeling
    - Drop caches
    - Drop persistent notifications database

    23 -> 24
    - Migrating article attachments to use smart IDs

    24 -> 6000
    - Migrating project path to project folder
    - Change project home behavior in user-settings.json
    - Migrating article IDs to numeric IDs instead of names
    - Migrating homepage articles to ids

    [Install] - Migrating dashboard articles to ids
    - Upgrade Wiki timeline with numeric article IDs
    - Change credentials mode in AWS connections
    - Upgrade pre-push hook in general settings
    - Upgrade pre-push hook in API deployer
    - Migrating visual ML execution params
    - Migrating listing of code-envs used in visual deep-learning
    - Upgrade evaluation recipe metrics and ouputs configurations

    6000 -> 6020
    - Migrating article IDs to numeric IDs in dashboards

    6020 -> 6030
    - Migrating Microsoft Teams scenarios reporters

    6030 -> 7000
    - Renaming connectionCredentials to credentials for each user
    - Convert expositions in infras of API deployer
    - Convert expositions in deployments of API deployer
    - Removing existing Hipchat Channels
    - Removing existing Hipchat Reporters
    - Removing existing Hipchat Integrations
    - Migrating algorithms params structure
    - Migrating posttrain computation params
    - Migrating SAML SP configuration

    7000 -> 7020
    - Migrate scoring recipe explanation params

    7020 -> 8000
    - Initialize audit settings
    - Initialize audit settings (API node)
    - Migrating SVM algorithms params structure
    - Migrating structure of 'proportion of features to sample'
    - Migrating prediction algorithms params structure to add range
    - Migrating grid length for non-searchable algos
    - Migrate code env selection

    8000 -> 8020
    - Migrating tree visualisations to take into account difference between class and sample weights
    - Rename 'hashSVDhashSize' field to 'hashSize' for text features
    - Migrating structure of global tags categories

    8999 -> 9000
    - Migrate categorical hashing method
    - Update preparation script FilterOnDateRange steps
    - Update preparation script FlagOnDateRange steps
    - Migrating of the jupyter storage path to project
    - Add parameters to evaluation recipes
    - Migrate API service params to use type parameter
    - Migrating API Deployer settings to be Deployer settings
    - Migrating time triggers
    - Migrate max depth for tree based models


  • PGuti
    PGuti Registered Posts: 5 ✭✭✭✭✭
    edited 3:49PM


    So when looking at this file, I saw it was only containing "null". So I deleted it and restarted.

    I had the problem for several other files. So I deleted all of them. If I had to guess, I would say it's related to the custom models I had (the big WARNINGS on the log).

    In case someone has the same issue as I did, here is the python code to generate all file names you want to remove:

    import os
    import json
    def load_json_file(filename):
        with open(filename) as f:
            data = json.load(f)
        return data
    mainpath = "YOURPATHHERE"
    projects = os.listdir(mainpath)
    for p in projects:
        project_path = os.path.join(mainpath,p, "analysis")
        if os.path.isdir(project_path):
            analyses = os.listdir(project_path)
            for analysis in analyses:
                ml_path = os.path.join(project_path, analysis, "ml")
                if os.path.isdir(ml_path):
                    models = os.listdir(ml_path)
                    for m in models:
                        final_path = os.path.join(ml_path,m, "params.json")
                        if os.path.isfile(final_path):
                            mydata = load_json_file(final_path)
                            if mydata is None:
                                print("rm " + final_path)

    The studio is now working!

    Thanks for the help!

  • Huixian
    Huixian Dataiker, Dataiku DSS Core Designer, Dataiku DSS Adv Designer, Registered Posts: 2 Dataiker

    This is great! Thank you @PGuti
    for the updates and thank you for sharing the additional information!

    Huixian Pan

Setup Info
      Help me…