Is there a way to set the encoding in SparkConf? I'm building a Java application with Spark that processes Arabic data. When I run it in the dev environment with Spark Master set to Local[*], data is processed correctly. However, when I prepare the JAR and submit it to Spark Cluster, data seems to need encoding.
I used:
--conf spark.driver.extraJavaOptions=-Dfile.encoding=UTF-8
as configuration in Spark submit, but didn't work.
OS: Windows 10
Java 1.8.0.131
Spark 2.1.0