Skip to content
View emsalcengiz's full-sized avatar
🫦
cat
🫦
cat

Block or report emsalcengiz

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 250 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
emsalcengiz/README.md

Hi there, I'm Emsal!

💁‍♀️✨ I'm an Associate Data Engineer at Metyis, building smart, scalable, and cloud-native data pipelines for a global client.

What I do

🌐 Turning raw data into insights using:
    💙 Azure (Functions, Data Lake, Blob Storage)
    🐍 PySpark for large-scale batch and stream processing
    ⚙️ Apache Spark, Flink, Beam for distributed computing
    📊 Power BI dashboards for data storytelling

⏱ Working with both:

  • ⚡ Real-time data processing (Spark Structured Streaming, Flink, Beam)
  • 🧱 Batch workflows for data aggregation, modeling, and reporting

🧰 Also familiar with:

  • 🐳 Docker for isolated and portable environments
  • ⛓️ Apache Airflow for orchestration and scheduling
  • 📂 Parquet, Delta Lake for optimized storage
  • 🔁 REST APIs, webhooks & automation workflows
  • 🚀 CI/CD pipelines for smooth and fast deployments

📚 Currently Exploring

🔬 Cloud architecture & cost-efficient design
📐 Advanced data modeling & orchestration strategies
📈 Data reliability, observability, and product-driven metrics

⚡ Fun fact

I love making data pipelines that work like magic

Pinned Loading

  1. data-normalize-with-etl-procesess data-normalize-with-etl-procesess Public

    I made various data normalization operations with python scripts. Target data in CSV format

    Python 3 1

  2. Etl_processing Etl_processing Public

    I find Apache Airflow very useful for ETL work. Here I transferred data from the source database(mysql) to the target database(postresql) and used the Airflow Bash Operator.

    Python 1 1

  3. filtering-process filtering-process Public

    You can do a lot of things with Apache Spark. What I've done here is to work with a static file and create a Batch ETL system.

    Python

  4. get_users get_users Public

    After ETL done by reading static data, an API is designed with flask_sqlalchemy, the purpose is to show the top five users

    Python

  5. random-data-generation random-data-generation Public

    Generating random data with Apache Kafka

    Python

  6. Apache-Beam-examples Apache-Beam-examples Public

    liked Apache Beam for streaming data transformations

    Python