Spark is an open source cluster computing system that aims to make data analytics fast — both fast to run and fast to write.
To run programs faster, Spark provides primitives for in-memory
cluster computing: your job can load data into memory and query it
repeatedly much more quickly than with disk-based systems like Hadoop
MapReduce.
To make programming faster, Spark provides clean, concise APIs in Scala, Java and Python. You can also use Spark interactively from the Scala and Python shells to rapidly query big datasets.
No comments:
Post a Comment