The equivalent of distributed caching in Spark?… here is a solution to the problem.
The equivalent of distributed caching in Spark?
In Hadoop, you can use a distributed cache to replicate read-only files on each node. What is the equivalent way to do this in Spark? I know about broadcast variables, but this only applies to variables, not files.