Hello, I am very new to Dataiku. My use-case might not be typical. I want to build an NLP pipeline, where each stage reads one file (say PDF) & produces another, say an XML file. I want to invoke a 3rd party program that requires 2 parameters, an HDFS input folder & an HDFS output folder. I see that a Recipe also requires an input & an output folder, for which I can define Managed folders. In my case, the actual Reading & Writing into HDFS folders will be performed by the 3rd party program, without using any DSS APIs. In that case, will I be hard-coding the folder names with-in my Recipe?
Here is the pseudo code for the Recipe:
HDFS_Input_Folder = "/input"
HDFS_Output_Folder = "/output"
#Now, where am I using the Managed Folders associated with the Recipe?
#The 3rd Party service is running on another server & Reads/Writes into HDFS folder
#I am using Dataiku only to build a Flow
#Can I use the information from the Managed Folder to build the "HDFS File Path" required by the 3rd Party app? How?