Skip to content

Apache Spark Installation

luongthevinh edited this page Jan 18, 2016 · 4 revisions

Apache Spark is a leading open-source solution for large-scale distributed in-memory computing.

Spark is most effective on large server clusters of several to several dozen machines. If you want to try out Spark on your personal computer, install it per the following instructions:

  • Go here;
  • For the "Choose a Spark release" option, select the latest version;
  • For the "Choose a package type" option, select "Pre-built for Hadoop 2.6 and later";
  • For the "Choose a download type" option, select "Direct Download";
  • Download the .tgz zipped file and unpack it; depending on the software you use, you may need to unpack multiple layers until you reach the true installation folder with sub-folders such as bin/ and lib/;
  • You may move the unpacked folder to another location you prefer.
Clone this wiki locally