🚀 Seamlessly run and scale Hadoop in Docker with zero hassle
-
Updated
Aug 28, 2025 - Dockerfile
🚀 Seamlessly run and scale Hadoop in Docker with zero hassle
Docker image builds for Hadoop sandbox.
HDFS local single node container for testing
Helm chart for Apache Hadoop using multi-arch docker images
🐳 hadoop ecosystems docker image
Hadoop3-HA-Docker is a production-ready, fault-tolerant Hadoop cluster deployed with Docker Compose. It automates the setup of a fully distributed Hadoop ecosystem with high availability (HA) features, designed for reliability, scalability, and real-world big data workloads
Local playground for Spark and Jupyter notebooks, plus Iceberg support
Set-up local spark cluster, hadoop (hdfs), airflow, postgresql on docker with ease, without any local installations
Apache Hadoop development environment integrated with Jupyter Notebook using Docker
Standalone Spark setup with Hadoop and Hive leveraged on docker containers.
A template repository provides convenient Apache Hadoop instance in Dev Containers.
A simple Big data stack with Docker
BigData Pipeline is a local testing environment for experimenting with various storage solutions (RDB, HDFS), query engines (Trino), schedulers (Airflow), and ETL/ELT tools (DBT). It supports MySQL, Hadoop, Hive, Kudu, and more.
A Docker image containing necessary tools for Big Data
Add a description, image, and links to the hadoop topic page so that developers can more easily learn about it.
To associate your repository with the hadoop topic, visit your repo's landing page and select "manage topics."