Mahmud Nabi

Hello, I'm

Mahmud Nabi

Senior Data Engineer

Turning data complexity into scalable business-impacting insights

Dhaka, Bangladesh
+880 155 246 53 31
Born 13/11/1998 Bangladeshi Citizen ✈️ Open to Relocate
Download Resume

About Me

I am a Senior Data Engineer with 6+ years of experience designing and building scalable data infrastructure. I specialize in turning complex data challenges into actionable business insights through modern data stack technologies.

Currently, I'm working at Optimizely, where I lead data engineering initiatives focusing on cloud data architecture, ELT pipeline optimization, and AI-powered data solutions. I'm passionate about data reliability, performance optimization, and implementing best practices like Test-Driven Development in data workflows.

Dhaka, Bangladesh munmridul@gmail.com +880 155 246 53 31

Technical Expertise

Languages

Python (6 years) SQL (5 years) PySpark (2 years) Bash

Data Warehousing

Snowflake (4 years) PostgreSQL (3 years) MySQL (2 years) Redshift MongoDB (2 years)

ELT & Orchestration

DBT (3 years) Airflow (4 years) Argo Workflow (3 years) Fivetran (3 years) Airbyte (3 years) Kafka (2 years)

Cloud & DevOps

AWS (4 years) Docker (5 years) Kubernetes (4 years) Terraform (2 years) GitHub Actions (3 years)

API & Development

FastAPI (3 years) Pydantic (3 years) REST APIs OOP/SOLID

AI & ML

MLflow (2 years) GraphRAG (1 years) Google ADK (1 years) Prompt Engineering (1 years) Scikit-learn

Work History

Senior Data Engineer

Optimizely | Dhaka, Bangladesh

06/2022 – Present

A US-based global enterprise empowering businesses to deliver optimized digital experiences at scale using AI-driven experimentation, personalization, and marketing orchestration.

Modern Data Stack Migration (AWS RDS Postgres to Snowflake)

Migration project to upgrade Optimizely's data warehouse to a modern cloud system (Snowflake), reducing cost and data reliability across the organization.

  • Migrating Airflow ELT pipelines to Argo Workflows on Kubernetes with dbt, enabling retries and backfills to achieve a 99.7% pipeline success rate
  • Deployed Airbyte on AWS EKS to replace Fivetran, cutting cloud costs by 50%
  • Designed Medallion architecture-based DBT pipelines enabling analytics across 30+ mission-critical business dashboards
  • Implemented Point-In-Time tables in DBT by leveraging DBT macros, eliminating ambiguity across different but relevant data by 100%
  • Implemented Terraform-based Snowflake RBAC for 2,000+ tables, enabling data security and governance with auditable access control
PythonSQLSnowflakeDBTAWS EKSAirbyteArgo WorkflowTerraform
DWBrainstormer (AI Agent) - AI-powered data query engine

AI-powered data query engine for general knowledge discovery in Optimizely's data.

  • Deployed a knowledge-graph-based AI agent for semantic data discovery, serving 40% of monthly recurring adhoc data requests
  • Built an Actor-Critique AI agent (using Google ADK) for automated Snowflake SQL generation, reducing manual query effort by 35-40%
  • Implemented agent-driven, schema-aware data quality tests with DBT, Argo, and GitHub Actions, reducing production incidents by 30%
PythonFastAPIDBTAWS EKSGraphRAGGoogle ADK
DataSync (Snowflake to Salesforce Reverse ETL)

Seamless integration of Salesforce and Snowflake, syncing customer insights from Snowflake to Salesforce to power smarter targeting.

  • Engineered a high-throughput Python-based Snowflake-Salesforce CRM integration, reliably ingesting up to 30 million rows per day
  • Refactored the legacy pipeline to incremental, idempotent processing, achieving atleast 50% reduction in cloud run costs
PythonSQLSnowflakeArgo WorkflowKubernetesTDD
ML Pipeline Infrastructure

Scalable ML pipeline infrastructure, reducing deployment time and accelerating Optimizely's ML-driven customer churn prediction analytics.

  • Optimized ML retraining by offloading pandas-based data transformations to Snowflake SQL, cutting runtime 60%
  • Scaled churn prediction analytics 5x by orchestrating 20+ ensemble ML models with memory-efficient, fault-tolerant Argo pipelines on AWS EKS
PythonAWSKubernetesMLflowScikit-learn

Software Engineer

Shohoz Limited | Dhaka, Bangladesh

01/2021 – 06/2022

One of the largest online ticketing platforms in Bangladesh.

PySpark-based ETL

Implemented PySpark data ingestion jobs for high-volume data processing.

  • Implemented PySpark data ingestion jobs (up to 30 GB/day from MySQL and MongoDB), optimizing joins and storage
  • Delivered analytics that cut go-to-market strategy review time by 3 days
PythonPySparkApache IcebergMySQLMongoDB

Featured Projects

Background

Education

Bachelor of Science, Computer Science and Engineering

Independent University Bangladesh, Dhaka, Bangladesh

01/2017 – 11/2021 | GPA: 3.43/4.00

Languages

English Fluent Bengali Native German A1

Hobbies

FPS Gaming Cars

Get In Touch

I'm currently open to new opportunities in Germany. Whether you have a question or just want to connect, feel free to reach out!

munmridul@gmail.com