Skip to content

Latest commit

 

History

History
76 lines (64 loc) · 3.01 KB

README.md

File metadata and controls

76 lines (64 loc) · 3.01 KB

Docker-Spark

Prerequisites

  • Install Docker Engine
  • Install docker-compose command

Configuration

Before starting the service, you must configure the configuration files to match your existing environment. Here are the configuration files that you need to prepare.

  • .env
  • conf/app.properties
  • conf/log4j2.properties
  • conf/spark-defaults.conf

Installation

1. Start Spark services and confirm the services are up.

sudo docker-compose up -d
sudo docker-compose ps
  • Result
     Name                    Command               State                        Ports                    
----------------------------------------------------------------------------------------------------------
spark-master     /opt/entrypoint.sh bash -c ...   Up      0.0.0.0:7077->7077/tcp,:::7077->7077/tcp, 0.0.0.0:8180->8080/tcp,:::8180->8080/tcp
spark-worker-1   /opt/entrypoint.sh bash -c ...   Up      0.0.0.0:8181->8081/tcp,:::8181->8081/tcp                                          
spark-worker-2   /opt/entrypoint.sh bash -c ...   Up      0.0.0.0:8182->8081/tcp,:::8182->8081/tcp                                          
spark-worker-3   /opt/entrypoint.sh bash -c ...   Up      0.0.0.0:8183->8081/tcp,:::8183->8081/tcp     

How to submit Spark Application

  1. SensorEnrichDataStreamJob
docker compose exec -it -w /opt/spark spark-master /opt/spark/bin/spark-submit \
     --master spark://spark-master:7077 \
     --class org.mataelang.kaspacore.jobs.SensorEnrichDataStreamJob \
     --total-executor-cores 1 \
     --conf spark.submit.deployMode=cluster \
     --conf spark.driver.memory=1g \
     --conf spark.executor.cores=1 \
     --conf spark.executor.memory=1g \
     --conf spark.eventLog.enabled=true \
     --conf spark.eventLog.dir=hdfs://192.168.1.121:9000/user/fadhilyori/spark/spark-events \
     --files conf/app.properties \
     hdfs://192.168.1.121:9000/user/fadhilyori/kaspacore/files/kaspacore.jar
  1. SensorAggregationStreamJob
docker compose exec -it -w /opt/spark spark-master /opt/spark/bin/spark-submit \
     --master spark://spark-master:7077 \
     --class org.mataelang.kaspacore.jobs.SensorAggregationStreamJob \
     --total-executor-cores 3 \
     --conf spark.submit.deployMode=client \
     --conf spark.driver.memory=1g \
     --conf spark.executor.cores=1 \
     --conf spark.executor.memory=1g \
     --conf spark.eventLog.enabled=true \
     --conf spark.eventLog.dir=hdfs://192.168.1.121:9000/user/fadhilyori/spark/spark-events \
     --conf spark.sql.shuffle.partitions=5 \
     --conf spark.sql.codegen.aggregate.map.twolevel.enabled=false \
     --conf spark.sql.streaming.metricsEnabled=true \
     --files conf/app.properties \
     hdfs://192.168.1.121:9000/user/fadhilyori/kaspacore/files/kaspacore.jar

beli Pi modul compute 4 + DeskPi Super6C udh cukup ini buat dibawa demo2 gak berat wkwk

https://www.raspberrypi.com/products/compute-module-4

https://deskpi.com/products/deskpi-super6c-raspberry-pi-cm4-cluster-mini-itx-board-6-rpi-cm4-supported