Software Developer · Data Engineer · Cloud Engineer · Data Analyst
Qualified Software Developer with expertise in building production-grade data pipelines, cloud infrastructure, and analytical solutions — using Python, AWS, SQL, Docker, and Airflow to deliver real-world impact.
I'm a Qualified Software Developer and AWS Certified Data Engineer based in Durban, South Africa — currently working at Africa Health Research Institute (AHRI), where I build and maintain production-grade ETL pipelines, Power BI dashboards, and clinical data systems that support real-world health research across Africa.
My work spans the full data lifecycle — from ingesting raw data and designing cloud infrastructure on AWS, to transforming, validating, and serving data insights to stakeholders. I'm passionate about building reliable, scalable systems that turn messy data into clear decisions.
Outside of data engineering, I enjoy exploring new AWS services, building side projects, and continuously levelling up — currently expanding into AWS AI services and real-time streaming pipelines.
📍 Location
Durban, KwaZulu-Natal, South Africa
💼 Current Role
Data Manager · Africa Health Research Institute
🎓 Education
PGDip Computer Science · Tshwane University of Technology
🏅 Certifications
AWS Cloud Practitioner · AWS Data Engineer Associate
Building scalable cloud infrastructure using AWS Free Tier services for real-world data engineering workloads.
Writing production-grade Python for data ingestion, transformation, and automated quality validation pipelines.
Designing and deploying automated DAGs with Apache Airflow for scheduled, monitored, and reliable data pipelines.
Implementing Medallion Architecture (Raw → Silver → Gold) for clean, scalable, and queryable data lake designs.
Querying and analysing large datasets using SQL on Athena and Python for delivery, sales, and e-commerce insights.
Containerising services with Docker Compose and managing code with Git and GitHub for version control and CI/CD.
Production-grade data engineering pipeline on AWS Free Tier. Ingests e-commerce data from Fake Store API, transforms JSON to Parquet via Medallion Architecture, catalogs with Glue, queries with Athena — fully orchestrated by Apache Airflow DAGs running in Docker.
Full delivery data pipeline project — ingesting, processing, and analysing delivery data through an automated data engineering workflow built for real-world logistics use cases.
In-depth analysis of delivery data — exploring patterns, performance metrics, and operational insights using Python and data analysis techniques to drive business decisions.
Designed and implemented a sales data warehouse — structured for efficient querying and reporting on sales performance, trends, and KPIs using modern data warehousing principles.
Analysis of facility visit data — uncovering usage patterns, peak times, and operational insights to support data-driven facility management and resource planning decisions.
Cloud infrastructure project deploying a Node.js application on AWS EC2 with Nginx as a reverse proxy — demonstrating cloud deployment, server configuration, and DevOps skills.
End-to-end data lake ingesting South African and African artist data from YouTube API, Last.fm, and MusicBrainz — featuring automated ETL scripts and a live Power BI dashboard.
Africa Health Research Institute (AHRI) · Full-time
Durban, KwaZulu-Natal · Hybrid
CSG · Internship
Centurion, Gauteng · Hybrid
Amazon Web Services
Data Engineering Certification
Computer Science
Computer Science
Computer Science