[2023/03/07-21:23:05.821] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.analysis.prediction] - [ct: 0] ****************************************** [2023/03/07-21:23:05.840] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.analysis.prediction] - [ct: 19] ** Start train session s18 [2023/03/07-21:23:05.862] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.analysis.prediction] - [ct: 41] ****************************************** [2023/03/07-21:23:05.893] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.analysis.splits] T-xC3fKPp5 - [ct: 72] Search for split: p=type=SPLIT_SINGLE_DATASET,split=RANDOM,splitBeforePrepare=true,ds=customers_labeled,sel=(method=head-s,records=100000),r=0.8,s=1337 i=b0ce29366428a118e7d4215ac78d8fe3-3 [2023/03/07-21:23:05.921] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.shaker.data] T-xC3fKPp5 - [ct: 100] Need to compute sampleId before checking memory cache [2023/03/07-21:23:05.940] [FT-TrainWorkThread-6vtxnj8S-445] [DEBUG] [dip.shaker.runner] T-xC3fKPp5 - [ct: 119] Script settings sampleMax=104857600 processedMax=-1 [2023/03/07-21:23:05.962] [FT-TrainWorkThread-6vtxnj8S-445] [DEBUG] [dip.shaker.runner] T-xC3fKPp5 - [ct: 141] Processing with sampleMax=104857600 processedMax=524288000 [2023/03/07-21:23:05.982] [FT-TrainWorkThread-6vtxnj8S-445] [DEBUG] [dip.shaker.runner] T-xC3fKPp5 - [ct: 161] Computed required sample id : 4f255aac20ac690dbbda55c2dd2236d3-NA-ac2e0aa81c1215a34ba9f85052ba5ff70--d751713988987e9331980363e24189ce [2023/03/07-21:23:06.002] [FT-TrainWorkThread-6vtxnj8S-445] [DEBUG] [dku.shaker.cache] T-xC3fKPp5 - Shaker MemoryCache get on dataset DKU_TUTORIAL_MACHINE_LEARNING_BASICS.customers_labeled key=ds=0c7519fdd1c1ecbf517d6cb46ea886d7--scr=6b1917bfb0eb4a902c208916dbbb80e8--samp=4f255aac20ac690dbbda55c2dd2236d3-NA-ac2e0aa81c1215a34ba9f85052ba5ff70--d751713988987e9331980363e24189ce: hit [2023/03/07-21:23:06.023] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.shaker.schema] T-xC3fKPp5 - [ct: 202] Column customerID meaning=Text fail=0 [2023/03/07-21:23:06.043] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.shaker.schema] T-xC3fKPp5 - [ct: 222] Column gender meaning=Gender fail=0 [2023/03/07-21:23:06.063] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.shaker.schema] T-xC3fKPp5 - [ct: 242] Column age_first_order meaning=LongMeaning fail=0 [2023/03/07-21:23:06.084] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.shaker.schema] T-xC3fKPp5 - [ct: 263] Column user_agent_brand meaning=Text fail=0 [2023/03/07-21:23:06.103] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.shaker.schema] T-xC3fKPp5 - [ct: 282] Column user_agent_os meaning=Text fail=0 [2023/03/07-21:23:06.125] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.shaker.schema] T-xC3fKPp5 - [ct: 304] Column ip_address_country meaning=CountryMeaning fail=0 [2023/03/07-21:23:06.148] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.shaker.schema] T-xC3fKPp5 - [ct: 327] Column ip_address_geopoint meaning=GeoPoint fail=0 [2023/03/07-21:23:06.169] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.shaker.schema] T-xC3fKPp5 - [ct: 348] Column campaign meaning=Boolean fail=0 [2023/03/07-21:23:06.190] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.shaker.schema] T-xC3fKPp5 - [ct: 369] Column pages_visited_avg meaning=DoubleMeaning fail=0 [2023/03/07-21:23:06.212] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.shaker.schema] T-xC3fKPp5 - [ct: 391] Column high_revenue meaning=Boolean fail=0 [2023/03/07-21:23:06.249] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.datasets.file] T-xC3fKPp5 - [ct: 428] Building Filesystem handler config: {"connection":"filesystem_managed","path":"/DKU_TUTORIAL_MACHINE_LEARNING_BASICS.customers_labeled","notReadyIfEmpty":false,"filesSelectionRules":{"mode":"ALL","excludeRules":[],"includeRules":[],"explicitFiles":[]}} [2023/03/07-21:23:06.269] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.datasets.ftplike] T-xC3fKPp5 - Enumerating Filesystem dataset prefix= [2023/03/07-21:23:06.287] [FT-TrainWorkThread-6vtxnj8S-445] [DEBUG] [dku.datasets.fsbased] T-xC3fKPp5 - [ct: 466] Building FS provider for dataset handler: DKU_TUTORIAL_MACHINE_LEARNING_BASICS.customers_labeled [2023/03/07-21:23:06.309] [FT-TrainWorkThread-6vtxnj8S-445] [DEBUG] [dku.datasets.fsbased] T-xC3fKPp5 - [ct: 488] FS Provider built [2023/03/07-21:23:06.331] [FT-TrainWorkThread-6vtxnj8S-445] [DEBUG] [dku.fs.local] T-xC3fKPp5 - [ct: 510] Enumerating local filesystem prefix=/ [2023/03/07-21:23:06.357] [FT-TrainWorkThread-6vtxnj8S-445] [DEBUG] [dku.fs.local] T-xC3fKPp5 - [ct: 536] Enumeration done nb_paths=8 size=186735 [2023/03/07-21:23:06.376] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.input.push] T-xC3fKPp5 - USTP: push selection.method=HEAD_SEQUENTIAL records=100000 ratio=0.02 col=null [2023/03/07-21:23:06.396] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.format] T-xC3fKPp5 - [ct: 575] Extractor run: limit={"maxBytes":-1,"maxRecords":100000,"ordering":{"enabled":false,"rules":[]}} totalRecords=0 [2023/03/07-21:23:06.416] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s0.csv.gz** [2023/03/07-21:23:06.435] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s0.csv.gz** [2023/03/07-21:23:06.454] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.format] T-xC3fKPp5 - [ct: 633] Start compressed [GZIP] stream: C:\Users\asd_v\AppData\Local\Dataiku\DataScienceStudio\dss_home\managed_datasets\DKU_TUTORIAL_MACHINE_LEARNING_BASICS.customers_labeled\out-s0.csv.gz / totalRecsBefore=0 [2023/03/07-21:23:06.473] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s0.csv.gz** [2023/03/07-21:23:06.494] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s0.csv.gz** [2023/03/07-21:23:06.523] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.format] T-xC3fKPp5 - [ct: 702] after stream totalComp=23065 totalUncomp=104367 totalRec=1348 [2023/03/07-21:23:06.547] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s1.csv.gz** [2023/03/07-21:23:06.567] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s1.csv.gz** [2023/03/07-21:23:06.589] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.format] T-xC3fKPp5 - [ct: 768] Start compressed [GZIP] stream: C:\Users\asd_v\AppData\Local\Dataiku\DataScienceStudio\dss_home\managed_datasets\DKU_TUTORIAL_MACHINE_LEARNING_BASICS.customers_labeled\out-s1.csv.gz / totalRecsBefore=1348 [2023/03/07-21:23:06.608] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s1.csv.gz** [2023/03/07-21:23:06.627] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s1.csv.gz** [2023/03/07-21:23:06.659] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.format] T-xC3fKPp5 - [ct: 838] after stream totalComp=46302 totalUncomp=208456 totalRec=2697 [2023/03/07-21:23:06.682] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s2.csv.gz** [2023/03/07-21:23:06.703] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s2.csv.gz** [2023/03/07-21:23:06.723] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.format] T-xC3fKPp5 - [ct: 902] Start compressed [GZIP] stream: C:\Users\asd_v\AppData\Local\Dataiku\DataScienceStudio\dss_home\managed_datasets\DKU_TUTORIAL_MACHINE_LEARNING_BASICS.customers_labeled\out-s2.csv.gz / totalRecsBefore=2697 [2023/03/07-21:23:06.743] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s2.csv.gz** [2023/03/07-21:23:06.765] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s2.csv.gz** [2023/03/07-21:23:06.797] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.format] T-xC3fKPp5 - [ct: 976] after stream totalComp=70041 totalUncomp=313090 totalRec=4046 [2023/03/07-21:23:06.817] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s3.csv.gz** [2023/03/07-21:23:06.838] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s3.csv.gz** [2023/03/07-21:23:06.858] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.format] T-xC3fKPp5 - [ct: 1037] Start compressed [GZIP] stream: C:\Users\asd_v\AppData\Local\Dataiku\DataScienceStudio\dss_home\managed_datasets\DKU_TUTORIAL_MACHINE_LEARNING_BASICS.customers_labeled\out-s3.csv.gz / totalRecsBefore=4046 [2023/03/07-21:23:06.880] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s3.csv.gz** [2023/03/07-21:23:06.899] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s3.csv.gz** [2023/03/07-21:23:06.932] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.format] T-xC3fKPp5 - [ct: 1111] after stream totalComp=93428 totalUncomp=417550 totalRec=5394 [2023/03/07-21:23:06.955] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s4.csv.gz** [2023/03/07-21:23:06.977] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s4.csv.gz** [2023/03/07-21:23:06.999] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.format] T-xC3fKPp5 - [ct: 1178] Start compressed [GZIP] stream: C:\Users\asd_v\AppData\Local\Dataiku\DataScienceStudio\dss_home\managed_datasets\DKU_TUTORIAL_MACHINE_LEARNING_BASICS.customers_labeled\out-s4.csv.gz / totalRecsBefore=5394 [2023/03/07-21:23:07.019] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s4.csv.gz** [2023/03/07-21:23:07.038] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s4.csv.gz** [2023/03/07-21:23:07.068] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.format] T-xC3fKPp5 - [ct: 1247] after stream totalComp=116716 totalUncomp=521343 totalRec=6742 [2023/03/07-21:23:07.087] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s5.csv.gz** [2023/03/07-21:23:07.111] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s5.csv.gz** [2023/03/07-21:23:07.131] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.format] T-xC3fKPp5 - [ct: 1310] Start compressed [GZIP] stream: C:\Users\asd_v\AppData\Local\Dataiku\DataScienceStudio\dss_home\managed_datasets\DKU_TUTORIAL_MACHINE_LEARNING_BASICS.customers_labeled\out-s5.csv.gz / totalRecsBefore=6742 [2023/03/07-21:23:07.153] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s5.csv.gz** [2023/03/07-21:23:07.176] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s5.csv.gz** [2023/03/07-21:23:07.207] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.format] T-xC3fKPp5 - [ct: 1386] after stream totalComp=140211 totalUncomp=625398 totalRec=8090 [2023/03/07-21:23:07.231] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s6.csv.gz** [2023/03/07-21:23:07.253] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s6.csv.gz** [2023/03/07-21:23:07.274] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.format] T-xC3fKPp5 - [ct: 1453] Start compressed [GZIP] stream: C:\Users\asd_v\AppData\Local\Dataiku\DataScienceStudio\dss_home\managed_datasets\DKU_TUTORIAL_MACHINE_LEARNING_BASICS.customers_labeled\out-s6.csv.gz / totalRecsBefore=8090 [2023/03/07-21:23:07.292] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s6.csv.gz** [2023/03/07-21:23:07.311] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s6.csv.gz** [2023/03/07-21:23:07.343] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.format] T-xC3fKPp5 - [ct: 1522] after stream totalComp=163264 totalUncomp=729569 totalRec=9438 [2023/03/07-21:23:07.364] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s7.csv.gz** [2023/03/07-21:23:07.382] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s7.csv.gz** [2023/03/07-21:23:07.401] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.format] T-xC3fKPp5 - [ct: 1580] Start compressed [GZIP] stream: C:\Users\asd_v\AppData\Local\Dataiku\DataScienceStudio\dss_home\managed_datasets\DKU_TUTORIAL_MACHINE_LEARNING_BASICS.customers_labeled\out-s7.csv.gz / totalRecsBefore=9438 [2023/03/07-21:23:07.421] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s7.csv.gz** [2023/03/07-21:23:07.440] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku] T-xC3fKPp5 - getCompression filename=**out-s7.csv.gz** [2023/03/07-21:23:07.472] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.format] T-xC3fKPp5 - [ct: 1651] after stream totalComp=186735 totalUncomp=833666 totalRec=10787 [2023/03/07-21:23:07.492] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.format] T-xC3fKPp5 - [ct: 1671] Extractor run done, totalCompressed=186735 totalRecords=10787 [2023/03/07-21:23:07.521] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.analysis.splits] T-xC3fKPp5 - [ct: 1700] Checking if splits are up to date. Policy: type=SPLIT_SINGLE_DATASET,split=RANDOM,splitBeforePrepare=true,ds=customers_labeled,sel=(method=head-s,records=100000),r=0.8,s=1337, instance id: b0ce29366428a118e7d4215ac78d8fe3-3 [2023/03/07-21:23:07.546] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.analysis.splits] T-xC3fKPp5 - [ct: 1725] Search for split: p=type=SPLIT_SINGLE_DATASET,split=RANDOM,splitBeforePrepare=true,ds=customers_labeled,sel=(method=head-s,records=100000),r=0.8,s=1337 i=b0ce29366428a118e7d4215ac78d8fe3-3 [2023/03/07-21:23:07.577] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.analysis.splits] T-xC3fKPp5 - [ct: 1756] Search for split: p=type=SPLIT_SINGLE_DATASET,split=RANDOM,splitBeforePrepare=true,ds=customers_labeled,sel=(method=head-s,records=100000),r=0.8,s=1337 i=b0ce29366428a118e7d4215ac78d8fe3-3 [2023/03/07-21:23:07.623] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.analysis.splits] T-xC3fKPp5 - [ct: 1802] Checking if splits are up to date. Policy: type=SPLIT_SINGLE_DATASET,split=RANDOM,splitBeforePrepare=true,ds=customers_labeled,sel=(method=head-s,records=100000),r=0.8,s=1337, instance id: b0ce29366428a118e7d4215ac78d8fe3-3 [2023/03/07-21:23:07.645] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.analysis.splits] T-xC3fKPp5 - [ct: 1824] Search for split: p=type=SPLIT_SINGLE_DATASET,split=RANDOM,splitBeforePrepare=true,ds=customers_labeled,sel=(method=head-s,records=100000),r=0.8,s=1337 i=b0ce29366428a118e7d4215ac78d8fe3-3 [2023/03/07-21:23:07.676] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.analysis.splits] T-xC3fKPp5 - [ct: 1855] Search for split: p=type=SPLIT_SINGLE_DATASET,split=RANDOM,splitBeforePrepare=true,ds=customers_labeled,sel=(method=head-s,records=100000),r=0.8,s=1337 i=b0ce29366428a118e7d4215ac78d8fe3-3 [2023/03/07-21:23:07.708] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.analysis.ml.python] T-xC3fKPp5 - [ct: 1887] Joining processing thread ... [2023/03/07-21:23:07.709] [MRT-447] [INFO] [dku.analysis.ml.python] - Running a preprocessing set: pp1 in C:\Users\asd_v\AppData\Local\Dataiku\DataScienceStudio\dss_home\analysis-data\DKU_TUTORIAL_MACHINE_LEARNING_BASICS\sxTOK8gK\xC3fKPp5\sessions\s18\pp1 [2023/03/07-21:23:07.750] [MRT-447] [INFO] [dku.block.link] - Started a socket on port 21759 [2023/03/07-21:23:07.769] [MRT-447] [INFO] [dku.ml.kernel] - Writing output of python-single-command-kernel to C:\Users\asd_v\AppData\Local\Dataiku\DataScienceStudio\dss_home\analysis-data\DKU_TUTORIAL_MACHINE_LEARNING_BASICS\sxTOK8gK\xC3fKPp5\sessions\s18\pp1\train.log [2023/03/07-21:23:07.788] [MRT-447] [INFO] [dku.code.envs.resolution] - Executing Python activity in builtin env [2023/03/07-21:23:07.816] [MRT-447] [INFO] [dku.code.projectLibs] - EXTERNAL LIBS FROM DKU_TUTORIAL_MACHINE_LEARNING_BASICS is {"gitReferences":{},"pythonPath":["python"],"rsrcPath":["R"],"importLibrariesFromProjects":[]} [2023/03/07-21:23:07.833] [MRT-447] [INFO] [dku.code.projectLibs] - chunkFolder is C:\Users\asd_v\AppData\Local\Dataiku\DataScienceStudio\dss_home\config\projects\DKU_TUTORIAL_MACHINE_LEARNING_BASICS\lib\R [2023/03/07-21:23:07.853] [MRT-447] [INFO] [dku.python.single_command.kernel] - Starting Python process for kernel python-single-command-kernel [2023/03/07-21:23:07.879] [MRT-447] [INFO] [dip.tickets] - Creating API ticket for analysis-ml-DKU_TUTORIAL_MACHINE_LEARNING_BASICS-PJ3gftC on behalf of admin id=analysis-ml-DKU_TUTORIAL_MACHINE_LEARNING_BASICS-PJ3gftC_B8ePoBN4MZpq [2023/03/07-21:23:07.888] [MRT-447] [INFO] [dku.security.process] - Starting process (regular) [2023/03/07-21:23:07.909] [MRT-447] [INFO] [dku.security.process] - Process started with pid=5344 [2023/03/07-21:23:07.929] [MRT-447] [INFO] [dku.processes.cgroups] - Will use cgroups [] [2023/03/07-21:23:07.959] [MRT-447] [INFO] [dku.processes.cgroups] - Applying rules to used cgroups: [] [2023/03/07-21:23:07.982] [KNL-python-single-command-kernel-monitor-456] [DEBUG] [dku.resourceusage] - Reporting start of CRU:{"context":{"type":"ANALYSIS_ML_TRAIN","authIdentifier":"admin","projectKey":"DKU_TUTORIAL_MACHINE_LEARNING_BASICS","analysisId":"sxTOK8gK","mlTaskId":"xC3fKPp5","sessionId":"s18"},"type":"LOCAL_PROCESS","id":"EchLtt3Pry2J3QEm","startTime":1678216987982,"localProcess":{"cpuCurrent":0.0}} [2023/03/07-21:23:08.004] [process-resource-monitor-5344-457] [DEBUG] [dku.resource] - Process stats for pid 5344: {"pid":5344,"commandName":"C:\\Users\\asd_v\\AppData\\Local\\Dataiku\\DataScienceStudio\\dss_home\\pyenv\\Scripts\\python.exe","cpuCurrent":0.0,"vmRSSTotalMBS":0} Traceback (most recent call last): File "C:\Users\asd_v\AppData\Local\Dataiku\DataScienceStudio\kits\dataiku-dss-11.3.1-win\pythonwin.packages\numpy\core\__init__.py", line 16, in from . import multiarray ImportError: DLL load failed: The file cannot be accessed by the system. During handling of the above exception, another exception occurred: Traceback (most recent call last): File "C:\Users\asd_v\AppData\Local\Dataiku\DataScienceStudio\Python\python-3.7.13\lib\runpy.py", line 183, in _run_module_as_main mod_name, mod_spec, code = _get_module_details(mod_name, _Error) File "C:\Users\asd_v\AppData\Local\Dataiku\DataScienceStudio\Python\python-3.7.13\lib\runpy.py", line 109, in _get_module_details __import__(pkg_name) File "C:\Users\asd_v\AppData\Local\Dataiku\DataScienceStudio\kits\dataiku-dss-11.3.1-win\python\dataiku\__init__.py", line 13, in from .core.dataset import Dataset, _dataset_writer_atexit_handler File "C:\Users\asd_v\AppData\Local\Dataiku\DataScienceStudio\kits\dataiku-dss-11.3.1-win\python\dataiku\core\dataset.py", line 8, in import numpy as np File "C:\Users\asd_v\AppData\Local\Dataiku\DataScienceStudio\kits\dataiku-dss-11.3.1-win\pythonwin.packages\numpy\__init__.py", line 142, in from . import add_newdocs File "C:\Users\asd_v\AppData\Local\Dataiku\DataScienceStudio\kits\dataiku-dss-11.3.1-win\pythonwin.packages\numpy\add_newdocs.py", line 13, in from numpy.lib import add_newdoc File "C:\Users\asd_v\AppData\Local\Dataiku\DataScienceStudio\kits\dataiku-dss-11.3.1-win\pythonwin.packages\numpy\lib\__init__.py", line 8, in from .type_check import * File "C:\Users\asd_v\AppData\Local\Dataiku\DataScienceStudio\kits\dataiku-dss-11.3.1-win\pythonwin.packages\numpy\lib\type_check.py", line 11, in import numpy.core.numeric as _nx File "C:\Users\asd_v\AppData\Local\Dataiku\DataScienceStudio\kits\dataiku-dss-11.3.1-win\pythonwin.packages\numpy\core\__init__.py", line 26, in raise ImportError(msg) ImportError: Importing the multiarray numpy extension module failed. Most likely you are trying to import a failed build of numpy. If you're working with a numpy git repo, try `git clean -xdf` (removes all files not under version control). Otherwise reinstall numpy. Original error was: DLL load failed: The file cannot be accessed by the system. [2023/03/07-21:23:08.158] [KNL-python-single-command-kernel-monitor-456] [INFO] [dku.kernels] - Process done with code 1 [2023/03/07-21:23:08.168] [KNL-python-single-command-kernel-monitor-456] [INFO] [dip.tickets] - Destroying API ticket for analysis-ml-DKU_TUTORIAL_MACHINE_LEARNING_BASICS-PJ3gftC on behalf of admin [2023/03/07-21:23:08.186] [KNL-python-single-command-kernel-monitor-456] [DEBUG] [dku.resourceusage] - Reporting completion of CRU:{"context":{"type":"ANALYSIS_ML_TRAIN","authIdentifier":"admin","projectKey":"DKU_TUTORIAL_MACHINE_LEARNING_BASICS","analysisId":"sxTOK8gK","mlTaskId":"xC3fKPp5","sessionId":"s18"},"type":"LOCAL_PROCESS","id":"EchLtt3Pry2J3QEm","startTime":1678216987982,"localProcess":{"pid":5344,"commandName":"C:\\Users\\asd_v\\AppData\\Local\\Dataiku\\DataScienceStudio\\dss_home\\pyenv\\Scripts\\python.exe","cpuCurrent":0.0,"vmRSSTotalMBS":0}} [2023/03/07-21:23:08.196] [KNL-python-single-command-kernel-monitor-456] [ERROR] [dku.kernels] - KernelMonitorThread done: Closing: ServerSocket[addr=0.0.0.0/0.0.0.0,localport=21759] [2023/03/07-21:23:08.217] [MRT-447] [WARN] [dku.analysis.ml.python] - Training failed com.dataiku.dip.io.SocketBlockLink$SecretKernelTimeoutException: Subprocess failed to connect, it probably crashed at startup. Check the logs. at com.dataiku.dip.io.SocketBlockLink.waitForConnection(SocketBlockLink.java:91) at com.dataiku.dip.io.SecretProtectedKernelLink.waitForProcess(SecretProtectedKernelLink.java:53) at com.dataiku.dip.io.SecretProtectedKernelLink.waitForProcess(SecretProtectedKernelLink.java:90) at com.dataiku.dip.io.PythonSecretProtectedKernel.start(PythonSecretProtectedKernel.java:122) at com.dataiku.dip.analysis.ml.shared.PRNSTrainThread.run(PRNSTrainThread.java:171) Caused by: java.net.SocketException: Socket operation on nonsocket: configureBlocking at java.net.DualStackPlainSocketImpl.configureBlocking(Native Method) at java.net.DualStackPlainSocketImpl.socketAccept(DualStackPlainSocketImpl.java:141) at java.net.AbstractPlainSocketImpl.accept(AbstractPlainSocketImpl.java:409) at java.net.PlainSocketImpl.accept(PlainSocketImpl.java:199) at java.net.ServerSocket.implAccept(ServerSocket.java:560) at java.net.ServerSocket.accept(ServerSocket.java:528) at com.dataiku.dip.io.SocketBlockLink.waitForConnection(SocketBlockLink.java:87) ... 4 more [2023/03/07-21:23:08.235] [MRT-447] [INFO] [dku.block.link] - Closed socket [2023/03/07-21:23:08.254] [MRT-447] [INFO] [dku.block.link] - Closed serverSocket [2023/03/07-21:23:08.263] [MRT-447] [ERROR] [dku.analysis.ml.python] - Processing failed com.dataiku.dip.io.SocketBlockLink$SecretKernelTimeoutException: Subprocess failed to connect, it probably crashed at startup. Check the logs. at com.dataiku.dip.io.SocketBlockLink.waitForConnection(SocketBlockLink.java:91) at com.dataiku.dip.io.SecretProtectedKernelLink.waitForProcess(SecretProtectedKernelLink.java:53) at com.dataiku.dip.io.SecretProtectedKernelLink.waitForProcess(SecretProtectedKernelLink.java:90) at com.dataiku.dip.io.PythonSecretProtectedKernel.start(PythonSecretProtectedKernel.java:122) at com.dataiku.dip.analysis.ml.shared.PRNSTrainThread.run(PRNSTrainThread.java:171) Caused by: java.net.SocketException: Socket operation on nonsocket: configureBlocking at java.net.DualStackPlainSocketImpl.configureBlocking(Native Method) at java.net.DualStackPlainSocketImpl.socketAccept(DualStackPlainSocketImpl.java:141) at java.net.AbstractPlainSocketImpl.accept(AbstractPlainSocketImpl.java:409) at java.net.PlainSocketImpl.accept(PlainSocketImpl.java:199) at java.net.ServerSocket.implAccept(ServerSocket.java:560) at java.net.ServerSocket.accept(ServerSocket.java:528) at com.dataiku.dip.io.SocketBlockLink.waitForConnection(SocketBlockLink.java:87) ... 4 more [2023/03/07-21:23:08.293] [MRT-447] [INFO] [dku.analysis.ml] - Locking model train info file C:\Users\asd_v\AppData\Local\Dataiku\DataScienceStudio\dss_home\analysis-data\DKU_TUTORIAL_MACHINE_LEARNING_BASICS\sxTOK8gK\xC3fKPp5\sessions\s18\pp1\m1\train_info.json [2023/03/07-21:23:08.307] [MRT-447] [INFO] [dku.analysis.ml] - Unlocking model train info file C:\Users\asd_v\AppData\Local\Dataiku\DataScienceStudio\dss_home\analysis-data\DKU_TUTORIAL_MACHINE_LEARNING_BASICS\sxTOK8gK\xC3fKPp5\sessions\s18\pp1\m1\train_info.json [2023/03/07-21:23:08.366] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.analysis.ml.python] T-xC3fKPp5 - [ct: 2545] Processing thread joined ... [2023/03/07-21:23:08.387] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.analysis.ml.python] T-xC3fKPp5 - [ct: 2566] Joining processing thread ... [2023/03/07-21:23:08.405] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.analysis.ml.python] T-xC3fKPp5 - [ct: 2584] Processing thread joined ... [2023/03/07-21:23:08.436] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.analysis] T-xC3fKPp5 - [ct: 2615] Train done [2023/03/07-21:23:08.463] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.analysis.prediction] T-xC3fKPp5 - [ct: 2642] Train done [2023/03/07-21:23:08.645] [FT-TrainWorkThread-6vtxnj8S-445] [INFO] [dku.analysis.trainingdetails] T-xC3fKPp5 - Publishing mltask-train-done reflected event