Replies: 1 comment 3 replies
-
It is not possible to share distributed objects like DataFrame between 2 different Spark applications, since an RDD/DataFrame is tied to the context that was used to create it. The only way to do this is to have the first application write your dataframe to a disk/storage account and then load it from the second application. |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I run Apache Spark as a standalone.
I have 2 Spark applications and these connected to Apache Spark with master URL:
The first application reads data from CSV and shares this data frame in some way
DataFrame dataFrame = spark.Read().Text("input.txt");
The second application needs to access this shared data frame.
How can I do this? Is it possible?
Beta Was this translation helpful? Give feedback.
All reactions