Using Jar based spark data source in Dataiku Recipe

DataikuNewUser1
Level 1
Using Jar based spark data source in Dataiku Recipe

Hi, 

I have a JAR file that i use as a spark datasource to read data from a url. 

This is how i do it on my local 

conf = SparkConf().setAll([("spark.jars", "path-to-jar/<jar_name>.jar"),])

df = spark.read.format("cloud.alpha.spark.providers.appobject.AppObjectTableProvider") \
    .option("appName", "<name>") \
    .option("appGroupName", "<name>") \
    .load()
But i am unable to replicate this on Dataiku recipe. I have kept the jar in managed folder, and when i am setting the path for the jar in spark config in the recipe the error is it is unable to find the specified path to the class: cloud.alpha.spark.providers.appobject.AppObjectTableProvider.
Can someone provide some help with this.
0 Kudos
3 Replies
SarinaS
Dataiker

Hi @DataikuNewUser1,

You can include additional Spark jar files under Administration > Settings > Spark > Advanced Settings > Additional jars:

Screenshot 2024-04-08 at 5.54.18 PM.png

 Can you see if this works for you?

Thank you,
Sarina

0 Kudos
DataikuNewUser1
Level 1
Author

Currently I am using a Demo community account. Is tis feature available there. If yes can u plz help me with the steps i can access this page

0 Kudos

Hi, 

If you don't have access to the administration you can't access this feature. 

I've added 2 more screenshots: 

Screenshot 2024-04-19 at 12.20.14 PM.png

Screenshot 2024-04-19 at 12.21.15 PM.png

Mattsco
0 Kudos