Ready for Dataiku 10? Try out the Crash Course on new features!GET STARTED

Installation / DSS Start failure

Solved!
PGuti
Level 2
Installation / DSS Start failure

Hello!

I had a quite old instance of DSS on my mac (I used it last time in late 2017).

I wanted to run it but I got an error when trying to launch it. 

So I thought OK I will just download the latest version and upgrade it. 

I did this and when I'm trying to start DSS I get the following error:

[Install] ******************************************
* MIGRATION FAILED
******************************************
* Attempting to rollback
Backing up partially-migrated: dss-version.json -> /Users/pierregutierrez/Library/DataScienceStudio/dss_home/_broken_migration_backup_20211029-103623/dss-version.json

[Install] Restoring backup: _pre_migration_backup_20211029-103623/dss-version.json -> dss-version.json

[Install] Backing up partially-migrated: bin/env-default.sh -> /Users/pierregutierrez/Library/DataScienceStudio/dss_home/_broken_migration_backup_20211029-103623/bin__env-default.sh
Restoring backup: _pre_migration_backup_20211029-103623/bin__env-default.sh -> bin/env-default.sh

[Install] Backing up partially-migrated: install.ini -> /Users/pierregutierrez/Library/DataScienceStudio/dss_home/_broken_migration_backup_20211029-103623/install.ini

[Install] Restoring backup: _pre_migration_backup_20211029-103623/install.ini -> install.ini
Backing up partially-migrated: databases -> /Users/pierregutierrez/Library/DataScienceStudio/dss_home/_broken_migration_backup_20211029-103623/databases

[Install] Restoring backup: _pre_migration_backup_20211029-103623/databases -> databases

[Install] Backing up partially-migrated: config -> /Users/pierregutierrez/Library/DataScienceStudio/dss_home/_broken_migration_backup_20211029-103623/config

[Install] Restoring backup: _pre_migration_backup_20211029-103623/config -> config
Restore complete, removing marker file

[Install] Traceback (most recent call last):
  File "/Applications/DataScienceStudio.app/Contents/Resources/kit/scripts/dkuinstall/migrate_auto.py", line 30, in <module>
    migrate_dss.migrate(sys.argv[1])
  File "/Applications/DataScienceStudio.app/Contents/Resources/kit/scripts/dkuinstall/migrate_dss.py", line 98, in migrate

[Install]     reraise(*sys.exc_info())
  File "/Applications/DataScienceStudio.app/Contents/Resources/kit/scripts/dkuinstall/migrate_dss.py", line 67, in migrate
    migration.execute(dip_home, node_type)
  File "/Applications/DataScienceStudio.app/Contents/Resources/kit/scripts/dkuinstall/migration_base.py", line 136, in execute

[Install]     wrapper.execute(diphome)
  File "/Applications/DataScienceStudio.app/Contents/Resources/kit/scripts/dkuinstall/migration_base.py", line 79, in execute
    migration.execute(ppaths)
  File "/Applications/DataScienceStudio.app/Contents/Resources/kit/scripts/dkuinstall/migration_json.py", line 132, in execute

[Install]     migrate_json_file(self, fp)
  File "/Applications/DataScienceStudio.app/Contents/Resources/kit/scripts/dkuinstall/migration_json.py", line 95, in migrate_json_file
    new_obj = transformer.transform(cobj, fp)
  File "/Applications/DataScienceStudio.app/Contents/Resources/kit/scripts/dkuinstall/steps.py", line 4940, in transform

[Install]     return migrate_ml_spark_params(modeling)
  File "/Applications/DataScienceStudio.app/Contents/Resources/kit/scripts/dkuinstall/steps.py", line 4907, in migrate_ml_spark_params

[Install]     spark_params[prop] = obj.get(prop, None)
AttributeError: 'NoneType' object has no attribute 'get'

[install] EOF reached
[install] EOF reached
DSS installer terminated with status 1
DSS installer failed , fatal error

Any idea how to fix this? 

Thank you! 

0 Kudos
1 Solution
HuixianP
Dataiker
Dataiker

Hi @PGuti,

Thank you for the information!

Can you please check if the params.json file under "/Users/pierregutierrez/Library/DataScienceStudio/dss_home/config/projects/VELIB/analysis/W9NhaPCz/ml/7iRWiMaj/" is null/empty. If so, can you please delete the file or move it outside of the DSS data directory and try again with the migration?

If the issue persists after that, please attach the full log to the chat you opened, so that we can review and investigate further.

Thank you! 

Regards,
Huixian Pan

View solution in original post

5 Replies
sergeyd
Dataiker
Dataiker

Hi @PGuti

Can you please send us a couple of dozens of lines before this error happened? We need to see the last step the migration was doing before failure. 

Thanks. 

 

0 Kudos
PGuti
Level 2
Author

The complete log would show all the project names on a public page which I would prefer avoiding.

The exact line before was:

"""

[Install] Migrating /Users/pierregutierrez/Library/DataScienceStudio/dss_home/config/projects/VELIB/analysis/W9NhaPCz/ml/7iRWiMaj/params.json

"""

The last step written was:

"""

20 -> 21

 -  Migrates spark params in saved models into a common object

  (Wrapped)

"""

I had a bunch of warning messages: 
[Install] ** WARNING: Unhandled algorithm: KMEANS

[Install] ** WARNING: Unhandled algorithm: EXTRA_TREES

[Install] ** WARNING: Unhandled algorithm: NEURAL_NETWORK

[Install] ** WARNING: Unhandled algorithm: PY_TWO_STEP

which I'm assume are models which changed and custom models.

 

The beginning of the log seems to explicit what is going to be done and is:

"""

Starting Data Science Studio.app
dssHome:/Users/pierregutierrez/Library/DataScienceStudio/dss_home dssBin:/Users/pierregutierrez/Library/DataScienceStudio/dss_home/bin/dss
Starting Java check
/Applications/DataScienceStudio.app/Contents/Resources/kit/scripts/_find-java.sh
[Install] [+] Using Java at /Library/Internet Plug-Ins/JavaAppletPlugin.plugin/Contents/Home/bin/java : java version "1.8.0_121"

[Install] /Library/Internet Plug-Ins/JavaAppletPlugin.plugin/Contents/Home/bin/java

[Java check] EOF reached
[Java check] EOF reached
Java check terminated with status 0
dss-version exists, doing stop-upgrade-start
Running DSS task for 'stop'
[DSS] dss: DSS supervisor is not running

[dss stop] EOF reached
[dss stop] EOF reached
[Install] *********************************************
* Dataiku DSS installer *
*********************************************

[Install] [+] /Users/pierregutierrez/Library/DataScienceStudio/dss_home already exists and is a DSS installation

[Install] [+] Saving installation log to /Users/pierregutierrez/Library/DataScienceStudio/dss_home/run/install.log

[Install] [+] Using Java at /Library/Internet Plug-Ins/JavaAppletPlugin.plugin/Contents/Home/bin/java : java version "1.8.0_121"

[Install] [+] Checking required dependencies

[Install] + Detected OS distribution : osx 11.6

[Install] + Checking required packages...

[Install] [+] Migrating Python environment

[Install] [+] Precompiling Dataiku Python code

[Install] [+] Precompiling Jupyter Python code

[Install] [+] Precompiling third-party Python 2.7 code

[Install] [+] Migrating data directory

[Install] Select how to migrate from 17 from avail: [8 -> 9, 9 -> 10, 10 -> 11, 11 -> 12, 12 -> 13, 13 -> 14, 14 -> 15, 15 -> 16, 16 -> 17, 17 -> 18, 18 -> 19, 19 -> 20, 20 -> 21, 21 -> 22, 22 -> 23, 23 -> 24, 24 -> 6000, 6000 -> 6020, 6020 -> 6030, 6030 -> 7000, 7000 -> 7020, 7020 -> 8000, 8000 -> 8020, 8999 -> 9000]
Executing the following migrations


17 -> 18
- Clean up the partitioning of SQL datasets
- Upgrade the y/n resize-to-fit setting of dashboard images to new multi-option setting
- Cleanup trained models parameters
- Nest feature generation parameters
- Drop caches
- Drop persistent notifications database

 


18 -> 19

 


19 -> 20
- Add params to ML recipes
- Transform comments insight to discussions insights
- Transform dashboard by changing comments insight to discussions insights
- Drop caches
- Drop persistent notifications database


[Install]


20 -> 21
- Migrates spark params in saved models into a common object
- Migrates spark params in analyses' models into a common object
- Migrates spark params in analyses' ml tasks into a common object
- Regroup spark params in recipes
- Drop caches
- Drop persistent notifications database

 


21 -> 22
- Give NPS Survey Settings to each user

 


22 -> 23
-
[Install] Setup metadata for code envs in Jupyter kernels' specs
- Upgrade structure of dss_schema_info in H2 databases
- Make CLASS_WEIGHT the default weighting strategy for classification saved models
- Sets a default Git URL whitelist for clones/pulls
- Use smartName in refresh charts steps
- Move ml task param skipExpensiveReports in modeling
- Drop caches
- Drop persistent notifications database

 


23 -> 24
- Migrating article attachments to use smart IDs

 


24 -> 6000
- Migrating project path to project folder
- Change project home behavior in user-settings.json
- Migrating article IDs to numeric IDs instead of names
- Migrating homepage articles to ids
[Install]

[Install] - Migrating dashboard articles to ids
- Upgrade Wiki timeline with numeric article IDs
- Change credentials mode in AWS connections
- Upgrade pre-push hook in general settings
- Upgrade pre-push hook in API deployer
- Migrating visual ML execution params
- Migrating listing of code-envs used in visual deep-learning
- Upgrade evaluation recipe metrics and ouputs configurations

 


6000 -> 6020
- Migrating article IDs to numeric IDs in dashboards

 


6020 -> 6030
- Migrating Microsoft Teams scenarios reporters

 


6030 -> 7000
- Renaming connectionCredentials to credentials for each user
- Convert expositions in infras of API deployer
- Convert expositions in deployments of API deployer
- Removing existing Hipchat Channels
- Removing existing Hipchat Reporters
- Removing existing Hipchat Integrations
- Migrating algorithms params structure
- Migrating posttrain computation params
- Migrating SAML SP configuration

 


7000 -> 7020
- Migrate scoring recipe explanation params

 


7020 -> 8000
- Initialize audit settings
- Initialize audit settings (API node)
- Migrating SVM algorithms params structure
- Migrating structure of 'proportion of features to sample'
- Migrating prediction algorithms params structure to add range
- Migrating grid length for non-searchable algos
- Migrate code env selection

 


8000 -> 8020
- Migrating tree visualisations to take into account difference between class and sample weights
- Rename 'hashSVDhashSize' field to 'hashSize' for text features
- Migrating structure of global tags categories

 


8999 -> 9000
- Migrate categorical hashing method
- Update preparation script FilterOnDateRange steps
- Update preparation script FlagOnDateRange steps
- Migrating of the jupyter storage path to project
[Install]
- Add parameters to evaluation recipes
- Migrate API service params to use type parameter
- Migrating API Deployer settings to be Deployer settings
- Migrating time triggers
- Migrate max depth for tree based models

 

"""

0 Kudos
HuixianP
Dataiker
Dataiker

Hi @PGuti,

Thank you for the information!

Can you please check if the params.json file under "/Users/pierregutierrez/Library/DataScienceStudio/dss_home/config/projects/VELIB/analysis/W9NhaPCz/ml/7iRWiMaj/" is null/empty. If so, can you please delete the file or move it outside of the DSS data directory and try again with the migration?

If the issue persists after that, please attach the full log to the chat you opened, so that we can review and investigate further.

Thank you! 

Regards,
Huixian Pan

View solution in original post

PGuti
Level 2
Author

Hello!

So when looking at this file, I saw it was only containing "null". So I deleted it and restarted. 

I had the problem for several other files. So I deleted all of them. If I had to guess, I would say it's related to the custom models I had (the big WARNINGS on the log). 

In case someone has the same issue as I did, here is the python code to generate all file names you want to remove:

import os
import json

def load_json_file(filename):
    with open(filename) as f:
        data = json.load(f)
    return data

mainpath = "YOURPATHHERE"

projects = os.listdir(mainpath)

for p in projects:
    project_path = os.path.join(mainpath,p, "analysis")
    if os.path.isdir(project_path):
        analyses = os.listdir(project_path)
        for analysis in analyses:
            ml_path = os.path.join(project_path, analysis, "ml")
            if os.path.isdir(ml_path):
                models = os.listdir(ml_path)
                for m in models:
                    final_path = os.path.join(ml_path,m, "params.json")
                    if os.path.isfile(final_path):
                        mydata = load_json_file(final_path)
                        if mydata is None:
                            print("rm " + final_path)

The studio is now working!

Thanks for the help!

HuixianP
Dataiker
Dataiker

This is great! Thank you @PGuti for the updates and thank you for sharing the additional information! 

Regards,
Huixian Pan

0 Kudos
A banner prompting to get Dataiku DSS