java.lang.nullpointerexception for any receipt

gautviv
gautviv Registered Posts: 1 ✭✭✭✭
edited July 16 in Using Dataiku

Hey,

I'm working on DSS 2.2.1. I needed to reboot my dataiu server, and restart dataiku. Then, all my receipt (Hive and Python) don't work. I have the famous: java.lang.nullpointerexception. Here the error;


java.lang.NullPointerException
at java.util.Hashtable.put(Hashtable.java:514)
at java.util.Hashtable.putAll(Hashtable.java:587)
at org.apache.hadoop.conf.CoreDefaultProperties.<init>(CoreDefaultProperties.java:76)
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
at java.lang.Class.newInstance(Class.java:374)
at org.apache.hadoop.conf.Configuration.getProperties(Configuration.java:2064)
at org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2272)
at org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2224)
at org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2141)
at org.apache.hadoop.conf.Configuration.get(Configuration.java:1081)
at org.apache.hadoop.fs.FileSystem.getDefaultUri(FileSystem.java:177)
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:169)
at com.dataiku.dctc.HadoopLoader.getFS(HadoopLoader.java:42)
at com.dataiku.dip.datasets.fs.HDFSDatasetHandler.loadFS(HDFSDatasetHandler.java:479)
at com.dataiku.dip.datasets.fs.HDFSDatasetHandler.enumerateFilesystem(HDFSDatasetHandler.java:149)
at com.dataiku.dip.datasets.fs.AbstractFSDatasetHandler.enumeratePartition(AbstractFSDatasetHandler.java:384)
at com.dataiku.dip.datasets.fs.AbstractFSDatasetHandler.getContentHash(AbstractFSDatasetHandler.java:482)
at com.dataiku.dip.dataflow.ComputableHashComputer.getCurrentContentHash(ComputableHashComputer.java:153)
at com.dataiku.dip.dataflow.ComputableHashComputer.getCurrentContentHash(ComputableHashComputer.java:102)
at com.dataiku.dip.dataflow.jobrunner.ActivityRunner.checkSourcesReady(ActivityRunner.java:337)
at com.dataiku.dip.dataflow.jobrunner.ActivityRunner.runActivity(ActivityRunner.java:434)
at com.dataiku.dip.dataflow.jobrunner.JobRunner.runActivity(JobRunner.java:105)
at com.dataiku.dip.dataflow.jobrunner.JobRunner.access$700(JobRunner.java:29)
at com.dataiku.dip.dataflow.jobrunner.JobRunner$ActivityExecutorThread.run(JobRunner.java:281)

When I start dss I have this message in the stdout


[dataiku@lcl-dataiku-soft-1 bin]$ ./dss start
Waiting for Data Science Studio supervisor to start ...
backend STARTING
hproxy BACKOFF Exited too quickly (process log may have details)
ipython STARTING
nginx STARTING

Do you hava any idea what's going on? Thank's in advance

Gautier

Tagged:

Answers

  • Jbelafa
    Jbelafa Dataiker Posts: 21 Dataiker
    Hi

    There's definitely a problem with your HDFS. You can test it with the following shell command "hdfs -dfs -ls"

    (if it's a local hadoop cluster you may need to restart your Hadoop services) .

    Joel
Setup Info
    Tags
      Help me…