Hi, I’m Mohammad Mahdi

About Me

Mohammad Mahdi Azadjalal

I’m a Senior Data Engineer who loves solving problems at the intersection of data and technology. For me, data engineering isn’t just about moving data from A to B — it’s about designing systems that are resilient, trustworthy, and meaningful.

Over the past years at Snapp, I’ve worked on everything from real-time analytics pipelines and change data capture (CDC) systems to data quality frameworks and metadata services with DataHub. These experiences taught me how to balance scalability, performance, and governance in fast-paced environments.

Outside of work, I enjoy experimenting with side projects like sleep monitoring tools and IoT sensor management platforms — because I believe every dataset has a story to tell.

My goal is simple: help organizations build data ecosystems they can trust to drive smarter decisions.

Resume

Education

2012-2014
Kurdistan University of Technology

MSc, Artificial Intelligence

Specialized in AI with a focus on trust-aware recommender systems, exploring how trust can enhance recommendation algorithms for more reliable and personalized suggestions.

Articles Published:

  1. A trust-aware recommendation method based on Pareto dominance and confidence concepts Read more
  2. Application of game theory techniques for improving trust based recommender systems in social networks Read more
2009-2011
Hamedan University of Technology

BSc, Computer Engineering

Developed a thesis on optical character recognition (OCR) for continuous Persian handwritten numbers, contributing to advancements in automated text recognition for Persian scripts.

2006-2009
Hamedan University of Technology

ASc, Computer Engineering

Focused on developing an accounting system, gaining a solid foundation in software development and its practical application in financial systems.

Experience

June 2022-Current
Snapp

Senior Data Engineer

Snapp is Iran’s leading vehicle-for-hire service, offering a platform for on-demand rides via mobile applications.

Key Contributions:

  • Developed a web application for streamlining pipeline creation, including Debezium connectors and batch data pipelines, reducing development time.
  • Implemented custom operators in Airflow for complex data transformations and pipeline management, ensuring high-quality data processing.
  • Maintained services for schema management and Clickhouse query recommendations, optimizing data storage and retrieval.
  • Designed Spring Batch services to handle large datasets with advanced error handling, retry mechanisms, and skip policies.
  • Utilized Spark and Flink for large-scale data processing and real-time analytics, integrating Kafka for CDC (Change Data Capture) pipelines.
  • Managed cloud services on OKD, using CI/CD practices and ArgoCD for optimized deployment workflows.
  • Developed several freelance applications, including a sleep monitoring web app and sensor management systems for factories.
  • Implemented data quality and metadata services integrated with DataHub, enhancing lineage tracking, schema validation, and property checks across pipelines, including SSAS and other sources.
September 2021-June 2022
Holding RAD

Data Team Lead

Holding RAD is a leading group of companies in the financial, economic, and technological sectors.

Key Contributions:

  • Designed and implemented a KAPA architecture for data pipelines, including sink and source connectors.
  • Developed pipeline architecture for gathering data from different cryptocurrency exchanges and resources using asynchronous Python applications.
  • Implemented an Airflow cluster in distributed mode with Celery executor, running tasks with multi-workers and task groups.
  • Set up a TimescaleDB cluster as a time-series database with distributed hypertables.
  • Deployed a Kafka cluster in the pipeline and used Kafka Connect for sending and filtering data to TimescaleDB.
  • Designed and implemented APIs using FastAPI and GraphQL to provide useful endpoints for other products.
March 2021-September 2021
Hamrah Avval

Senior Data Engineer

Hamrah Avval, Iran's first and largest mobile network operator, offers diverse mobile communication services.

Key Contributions:

  • Developed automated bots for normalizing MongoDB collections and integrating data from Firebase into the Shad application.
  • Designed and implemented APIs using FastAPI for monitoring and managing key aspects of the Shad application, optimizing performance and data handling.
May 2017-March 2021
LifeWeb

Senior Data Engineer

LifeWeb is an AI-driven platform focused on social network and news media analysis, providing tools for business and organizational monitoring.

Key Contributions:

  • Developed big data pipelines based on lambda architecture using Kafka, Apache Spark, and Hadoop for real-time NLP processing (topic detection, NER, sentiment analysis) on social media data.
  • Designed batch data pipelines using Airflow, integrating data from Elasticsearch and applying advanced NLP algorithms like copy detection and user clustering.
  • Built a PySpark pipeline for data ingestion and transformation across multiple data resources.
  • Developed and maintained a web application using PHP/Laravel for managing social media post tagging, integrating with NLP algorithms.
  • Implemented DevOps processes, including Dockerizing GPU drivers and TensorFlow on Linux OS.
  • Led the development of a cryptocurrency investment management website as a freelance project, using Laravel and MySQL.
July 2015-May 2017
Arsham Machine Farzam

Software Developer

Arsham Machine Farzam specializes in the design and production of industrial machines, offering innovative solutions in automation and software development.

Key Contributions:

  • Implemented track-and-trace solutions for pharmaceutical and tobacco industries to ensure product authenticity, using C++ with Qt and PostgreSQL for backend support.
  • Developed a dashboard with Laravel for managing cigarette production machines and their operations.
  • Led embedded system designs and customer training on their usage, enhancing operational efficiency.
  • Developed and maintained the company's website, featuring product catalogs and manuals.
  • Conducted market research and software analysis to identify new product opportunities.
April 2008-July 2015
Freelancer

Software Developer

As a freelancer, I developed a variety of applications in different technologies and domains, working with clients to deliver efficient and innovative software solutions.

Key Contributions:

  • Developed a C# application for interfacing with the Pathology Sysmex machine to transmit patient data to a Hospital Information System (HIS) and display results.
  • Built a foot pressure detection application using C# and image processing to design 3D insoles for alleviating back and leg pain.
  • Created multiple Delphi 7 applications with SQL Server databases for companies, including an office register letters system and organizational rules management.
  • Designed and implemented multimedia advertising applications using Adobe Flash MX, distributed across various national platforms.

Data Engineering Expertise

Kafka and CDC Pipelines

85%

ETL and ELT Processes

85%

Airflow for Batch and Real-time Pipelines

80%

Data Lake Design with Clickhouse

70%

Metadata Management with DataHub

75%

Programming Skills

Python

95%

Java | Spring Boot | Spring Batch

70%

PHP

95%

C | C++ | Qt

85%

Delphi

95%

C#

50%

HTML | CSS

100%

Javascript

70%

System Design and Architecture

System Architecture Design

80%

Data Pipeline Design

90%

Big Data Architecture

70%

Microservices Architecture

50%

DevOps and Cloud Expertise

Docker and Kubernetes

60%

CI/CD Pipelines

75%

Cloud Services (OKD)

50%

Performance and Monitoring

Performance Tuning and Optimization

85%

Monitoring with Grafana and Prometheus

85%

Query Optimization

80%

Knowledges

  • Python
  • Java
  • Spring Boot
  • Spring Batch
  • FastAPI
  • MySQL
  • PostgreSQL
  • Spark
  • HDFS
  • Recommender Systems
  • Image Processing

Certificates

Portfolio

Mastering CDC with PostgreSQL, Debezium, and Kafka Connect: Challenges & Solutions

Mastering CDC with PostgreSQL, Debezium, and Kafka Connect: Challenges & Solutions

Medium
Preparing PostgreSQL for CDC: Essential Prerequisites for Streaming to Kafka with Debezium

Preparing PostgreSQL for CDC: Essential Prerequisites for Streaming to Kafka with Debezium

Medium
Powering Real-Time Data Pipelines with PostgreSQL: A Deep Dive into WAL and Logical Replication

Powering Real-Time Data Pipelines with PostgreSQL: A Deep Dive into WAL and Logical Replication

Medium

Contact

Iran, Tehran

mohammadmahdi.azadjalal@snapp.cab

mm.azadjalal@gmail.com

Ready to collaborate.

How can I assist you?