


Every project will have its own metastore and warehouse.Single Project Access (Single Project Single Connection).You can follow any of the three modes depending on your specific use-case. I have divided this article into three parts. What to ExpectĪt the end of this article, you should be able to create/run your Spark SQL projects and spark-shell on Windows OS. You can refer to the Scala project used in this article from GitHub here. Just make sure you'll downloading the correct OS-version from Spark's website. This article can also be used for setting up a Spark development environment on Mac or Linux as well. īy default, Spark SQL projects do not run on Windows OS and require us to perform some basic setup first that’s all we are going to discuss in this article, as I didn’t find it well documented anywhere over the internet or in books. It integrates easily with HIVE and HDFS and provides a seamless experience of parallel data processing. It provides implicit data parallelism and default fault tolerance. Now, this article is all about configuring a local development environment for Apache Spark on Windows OS.Īpache Spark is the most popular cluster computing technology, designed for fast and reliable computation. In my last article, I have covered how to set up and use Hadoop on Windows.
