You cannot edit this Postr after publishing. Are you sure you want to Publish?
Experience reading like never before
Sign in to continue reading.
Discover and read thousands of books from independent authors across India
Visit the bookstore"It was a wonderful experience interacting with you and appreciate the way you have planned and executed the whole publication process within the agreed timelines.”
Subrat SaurabhAuthor of Kuch Woh PalApache Spark is a unified computing engine and a set of libraries for parallel data processing on computer clusters.
Spark can be used from Python, Java, or Scala, R, or SQL. Spark itself is written in Scala, and runs on the Java Virtual Machine (JVM) and therefore to run Spark either on your laptop or a cluster, all you need is an installation of Java 6 or newer. If you wish to use the Python API you will also need a Python interpreter (version 2.6 or newer). If you wish to use R you’ll also need a version of R on your machine.
Downloading Spark for your Hadoop Version: You don’t need to have Hadoop, but if you have an existing Hadoop cluster or HDFS installation, download the matching version.
In this book, you will see Workouts of Spark using Scala version 2.11.12 as Spark is primarily written in Scala, making it Spark’s “default” language.
Suripeddi Koundinya
Suripeddi Koundinya is a Data Analyst, Blogger, Financial Advisor and Psychologist. He pursued Masters in Biotechnology, Astrology and Psychology. He authored books on Data Science – Hive, Scala, SQL, Big Data, Sqoop, etc. He’s both Fictional and Non-Fictional Writer.
The items in your Cart will be deleted, click ok to proceed.