Spark was designed to improve processing speeds for data analysis and manipulation programs.
It was written in Java and Scala and provides features not found in other systems, mostly because they're not mainstream nor that useful for non-data processing applications.
Here are some key features of "Spark":
· Programmatic APIs for developers
· In-memory data querying
· Interactive data mining
· Iterative algorithms
What's New in This Release: [ read full changelog ]
· Spark 0.7 adds a Python API called PySpark.
· Spark jobs now launch a web dashboard for monitoring the memory usage of each distributed dataset (RDD) in the program.
· Spark can now be built using Maven in addition to SBT.