Summary
Overview
Work History
Education
Skills
Timeline
Generic

Abdul Aziz

Lombard

Summary

Experienced Data Engineer with expertise in building and optimizing data pipelines using AWS (EC2, S3, Lambda, RDS, Glue), and Databricks for distributed data processing. Skilled in designing scalable ETL workflows, and automating data transformation with AWS Glue and Lambda. Proficient in integrating Tableau and Power BI to deliver interactive dashboards and insights for business stakeholders. Experienced in implementing CI/CD pipelines with Azure DevOps for seamless deployment and continuous integration. Strong background in database management with Amazon RDS and Redshift, focused on enabling data-driven decision-making, and improving operational efficiency.

Overview

8
8
years of professional experience

Work History

Data Engineer

Concord Systems Usa
Bolingbrook
12.2021 - Current
1. Data Pipeline Development & ETL
  • Designed, developed, and optimized end-to-end ETL pipelines for data extraction, transformation, and loading from multiple data sources into Amazon Redshift, S3, and RDS.
  • Developed scalable data workflows using AWS Lambda and AWS Glue to automate data processing and ensure smooth data integration across various platforms.
  • Utilized Databricks for building high-performance ETL pipelines, running distributed data processing workloads, and optimizing Spark jobs for large-scale data transformation.
  • Integrated Amazon S3 with other AWS services for storing large datasets and leveraging AWS Glue for managing metadata and schema evolution.
2. Cloud Data Architecture & AWS Services
  • Architected and deployed scalable data architectures using AWS services like EC2, FSx, RDS, S3, and CloudFront to store, process, and distribute data at scale.
  • Utilized Amazon RDS for relational database management and implemented AWS FSx for high-performance file systems, ensuring data processing efficiency.
  • Optimized data access and distribution through AWS CloudFront, enabling low-latency content delivery to geographically distributed users.
3. Data Quality, Monitoring, & Optimization
  • Established data quality checks and implemented monitoring in AWS CloudWatch to ensure the integrity and transparency of data pipelines.
  • Optimized SQL queries and ETL jobs for performance and resource efficiency, reducing processing times by X% using Databricks and AWS Lambda for serverless data operations.
  • Utilized AWS Glue for metadata management and automated data cleaning, reducing errors in the data pipeline and ensuring high-quality data availability for analytics.
4. Automation & Scripting
  • Developed Python scripts to automate data extraction, transformation, and loading, integrating with AWS Lambda and S3 to ensure seamless data flow and minimize manual intervention.
  • Automated data validation processes and pipeline monitoring, leveraging AWS CloudWatch for proactive issue resolution and ensuring system uptime.
5. Collaboration & Cross-Functional Work
  • Collaborated with data scientists, analysts, and business stakeholders to understand data requirements, developing optimized data solutions using AWS and Databricks.
  • Worked closely with DevOps teams to implement CI/CD pipelines using Azure DevOps, enabling automated deployment and continuous integration for data engineering workflows.
6. Data Analytics & Visualization
  • Designed and developed interactive dashboards and reports using Tableau and Power BI, providing stakeholders with real-time insights into key business metrics and performance.
  • Integrated Databricks with Tableau for advanced analytics and visualizations, allowing data scientists and analysts to access processed data directly from the Databricks environment.
  • Leveraged Power BI to create custom business intelligence reports and data visualizations, allowing non-technical users to explore and interpret complex data without relying on technical teams.
7. Big Data Technologies
  • Built and maintained data pipelines for processing structured and unstructured data from various sources, using Databricks for distributed data processing.
  • Utilized Apache Spark on Databricks for high-performance data processing, significantly improving data processing times for large datasets.
8. Database Management & Optimization
  • Proficient in Amazon RDS for managing relational databases, ensuring performance, backup, and scalability based on workload demands.
  • Managed and optimized data stored in MongoDB, designing data models and optimizing queries for efficient data retrieval in AWS environments.
9. Data Security & Compliance
  • Implemented data encryption protocols for secure data storage and transfer using AWS KMS, ensuring compliance with data security and privacy standards, such as GDPR and HIPAA.
  • Managed IAM roles and permissions to enforce data governance and ensure secure data access across teams and systems.
10. CI/CD & Automation
  • Implemented CI/CD pipelines using Azure DevOps, automating testing, building, and deployment of data engineering solutions, and ensuring continuous delivery across development, staging, and production environments.
  • Integrated Azure DevOps with AWS infrastructure for seamless deployment of data engineering projects, ensuring rapid release cycles, and high-quality data products.

Data Analyst

Vivvan Techno solutions
Hyderabad
02.2017 - 06.2019
  • Gained a strong understanding of core business processes and their impact on data workflows.
  • Developed and enforced coding standards, procedural guidelines, and change control protocols for application development in proprietary environments.
  • Led data quality initiatives and profiling tasks, ensuring data integrity from source systems.
  • Designed and implemented ETL processes, identifying source systems, mapping source-to-target relationships, cleansing data, and creating source specifications and ETL design documents. Followed best practices to ensure consistency and efficiency.
  • Managed data extraction, transformation, and loading (ETL) from diverse source systems, ensuring smooth data flow into target environments.
  • Designed optimized reporting tables to improve performance of complex queries and views.
  • Utilized external tables to transform and load data from legacy systems into target tables, streamlining data migration processes.
  • Implemented incremental loading of fact tables from source systems to staging tables on a daily basis to maintain data freshness.
  • Led the design and implementation of proof-of-concept solutions in the development environment, testing for feasibility and performance.
  • Expertise in SQL query optimization, database tuning, and performance tuning, including memory allocation and the use of various tools (e.g., export, import, load, autoloader, and database snapshots).
  • Provided day-to-day support for user queries, troubleshooting database issues, and ensuring smooth database operations.
  • Collaborated with user departments to gather requirements and ensure the efficient and effective use of data for business insights.

Education

Master's Information Technology -

Lindsey Wilson College
Columbia, KY

Bachelor's of Science -

Osmania University
Hyderabad, India

Skills

  • Cloud Services: AWS (S3, EC2, Lambda, RDS), Azure
  • CI/CD Tools: Azure DevOps (CI/CD Pipelines, Repos, Artifacts)
  • Data Engineering: Apache Airflow, Databricks
  • Data Governance: Immuta, Data Security, and Compliance
  • Data Ingestion: Data Lake Ingestion, ETL Process Automation
  • Business Intelligence: Tableau, Data Analysis, and Reporting
  • Programming & Scripting: Python, SQL
  • Version Control: Git, GitHub

Timeline

Data Engineer

Concord Systems Usa
12.2021 - Current

Data Analyst

Vivvan Techno solutions
02.2017 - 06.2019

Master's Information Technology -

Lindsey Wilson College

Bachelor's of Science -

Osmania University
Abdul Aziz