Alternatively, use the default location by clicking install now.
Python is another requirement for using Apache Spark on your Windows 10 PC. Once downloaded, you just need to double-click it and install it to your preferred location.Hit the Agree and Start Free Download button and specify the path you want to download the Java setup. Click on the Java Download button to go to the downloads section.In case you don’t have java installed on your PC, you can install it from the official Java website. It will display the installed java version on your PC.To check, you can type the command below and hit enter: Ensure that you have the latest or at least Java 8 on your PC. Installing JavaĪpache Spark requires at least Java 8 to install.
A Step-by-Step Guide to Install Apache Spark on Windows 10 Step 1. Once you make sure that you have all the prerequisites, you can proceed with installing Apache Spark on Windows 10.
Believe us, by the end of this article you will know how easy it is to install Apache Spark as this article will discuss the easy step-by-step guide on how to install Apache Spark on Windows 10. Once we are done with the installation of Spark and Scala, configure environment variable for SCALA_HOME and HADOOP_HOME.Some people think that installing Apache Spark on your PC is rocket science. Scala installed at "C:\Program Files (x86)\scala".
Note:- It is also possible to download source code and build using Maven or SBT.Refer this for other options of download.ĭownload Scala executables and install it.It is prerequisite for working with Apache spark, spark is written in Scala. I have placed it in following location E:\spark-1.5.2-bin-hadoop2.6. In this post we will setup spark and execute some sparks API's.ĭownload pre-build version of Apache spark and unzip it in some directory. For leaning Apache spark, it is very possible to setup it in standalone mode and start executing spark API's in Scala,Python or R shell. It is very possible to use spark with Hadoop HDFS, Amazon EC2 and others persistence storage system including local file system. Apache Spark is a general-purpose cluster computing system to process big data workloads.