Scala vs. Python for Apache Spark

Apache Spark is a great choice for cluster computing and includes language APIs for Scala, Java, Python, and R. Apache Spark includes libraries for SQL, streaming, machine learning, and graph processing. This broad set of functionality leads many developers to start developing against Apache Spark for distributed applications.   

The first big decision you have is where to run it. For most, that's a no-brainer - run it on Apache Hadoop YARN on your existing cluster. After that tough...

What feeling does this article give you?

#hashtags to follow:

Apache Spark [+]    APIs [+]    Scala [+]    Java [+]    Python [+]    SQL [+]    Apache Hadoop YARN [+]   

More #news: