Skip to content

AlanRMA/Pyspark-simple-docker-cluster

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Project overview

This project is designed to run on containers using Docker that can be expanded for Kubernetes usage in the future.

Getting Started

All you need to run your cluster is having Docker installed and follow Steps on terminal/shell.

Step-by-Step

0. Open Docker

1. Enter on you terminal

2. Go to project root folder

3. Run docker-compose up

4. Go to Containers in Docker

alt text

5. Select Master node

6. Go to Exec tab

alt text

7. At this point you can edit your task.py code to choose what you want to submit to cluster process

8. Run /opt/bitnami/spark/bin/spark-submit /opt/spark-app/task.py

Now you should see output on Exec tab from docker.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages