Skip to content

hskundhal/spark-docker-setup

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Environment

Unless you already have a working Apache Spark cluster, you will need to have Docker for simple environment setup.

The provided docker-compose.yml and Spark configurations in conf directory are cloned from https://github.com/gettyimages/docker-spark.

Setup

  1. Make sure Docker is installed properly and docker-compose is ready to use
  2. Run $ docker-compose up -d under the data-mr directory
  3. Check Spark UI at http://localhost:8080 and you should see 1 master and 1 worker
  4. Run $ docker exec -it datamr_master_1 /bin/bash to get into the container shell, and start utilizing Spark commands such as # spark-shell, # pyspark or # spark-submit. You may want to replace datamr_master_1 with actual container name that's spawned by the docker-compose process

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published