I've been searching for a while if there is any way to use a Scala class in Pyspark, and I haven't found any documentation nor guide about this subject. Let's
I am using IntelliJ 2019.1.3 Community Edition. In Scala compile server, JVM maximum heap size: 4096 my idea.vmoptions: -Xms4096m -Xmx6144m -XX:ReservedCodeCac