For my case, it looks like below once I set-up the environment variables - 4. Please try to keep the same folder structure. Based on what I have chosen, I will need to add the following variables as Environment variables - SPARK_HOME - C:\Spark\spark-2.2.1-bin-hadoop2.7 HADOOP_HOME - C:\Hadoop JAVA_HOME - C:\Program Files\Java\jdk1.8.0_191 These values are as per my folder structure. And Create New or Edit if already available. Let's set up the environment variable now. We will need all the above 3 Folder names in our next steps. It might be different based on what folder you choose. Typically it is like - C:\Program Files\Java\jdk1.8.0_191. Ideally Java version 8 works fine without any issues so far. If Java is not already installed, install it from Oracle website (). For Java - Check where your Java JDK is installed. But for this post, I am considering the C Drive for the set-up. Create and Verify The Folders:Ĭreate the below folders in C drive. Please do the following step by step and hopefully it should work for you - 1. We will also see some of the common errors people face while doing the set-up. This post explains How To Set up Apache Spark & PySpark in Windows 10. How To Set up Apache Spark & PySpark in Windows 10 ?
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |