Profilbild von HafizRanaSherHassanAli Ibrahim Experienced Data Engineer with Big Data Expertise aus Berlin

Hafiz Rana Sher Hassan Ali Ibrahim

verfügbar

Letztes Update: 19.03.2025

Experienced Data Engineer with Big Data Expertise

Abschluss: Bachelor in Computer Science
Stunden-/Tagessatz: anzeigen
Sprachkenntnisse: Englisch (verhandlungssicher)

Schlagwörter

Big Data Data Analysis Datenverarbeitung Cloud Computing ETL Data Warehousing Workflows Daten-Pipeline Airflow Amazon Web Services + 23 weitere Schlagwörter anzeigen

Dateianlagen

Hassan-Ali-Resume_250225.pdf

Skills

Data Engineering
Expertise in designing and implementing scalable data pipelines, data lakes, and data warehousing solutions using various big data technologies.

Cloud Platforms
Proficiency in working with major cloud platforms including AWS, Azure, and Google Cloud Platform for data processing and analytics.

Big Data Technologies
Advanced knowledge of Apache Spark, Hadoop, Kafka, and other big data frameworks for large-scale data processing and analysis.

Data Modeling
Skills in dimensional modeling, DBT, and cloud data warehousing in Snowflake, optimizing data architectures for analytics.

ETL & Data Pipeline Orchestration
Experience with Airflow, SSIS, and other tools for building and managing complex data workflows and ETL processes.

Programming & DevOps
Proficiency in Python and SQL for data processing, along with Docker, GIT, and Jenkins for CI/CD workflows.

Data Visualization
Capability to create dashboards and reports using tools like Power BI and Apache Superset for data-driven insights.
 

Projekthistorie

06/2022 - bis jetzt
Sr Data Engineer/Analyst
The Adecco Group (Marketing, PR und Design, 5000-10.000 Mitarbeiter)

Data engineering with the ability to design, implement, and deliver maintainable, high-quality code in Python/PySpark.
• Design and build data infrastructure, data plateform and delta lakehouse in Databricks.
• Developed and Optimized Compass, an internal Adecco product, to analyze supply and demand in the job market.
• Accomplished the migration of legacy Azure blob storage file system data lake to Databricks delta lake.
• Loaded data from various sources into delta lake, leveraging my expertise in data processing and aggregation to generate final delta tables. Loaded final data into CosmosDB using a document DB to enhance application performance. • Implemented seamless orchestration of Pyspark scripts through workflow in Databricks.
• Use Airflow to orchestrate the old legacy scripts.
• Build data normalization and company name matching algorithm to match internal and external companies.
• Built data models in dbt, implemented singular and generic tests to validate data integrity, and automated the execution of these tests within CI/CD pipelines using GitHub Actions.
• Responsible for peer code review and migration process.
• Conducted knowledge-sharing sessions with team members to foster collaboration and enhance overall team expertise.
• Implemented radius search functionality in Compass, an internal data product, using Geopandas, enhancing its capabilities to analyze and visualize location-based insights.
• Designed and implemented performance-driven metrics, including CPA (Cost Per Application) and CPSH (Cost Per Successful Hire), to optimize marketing spend and enhance ROI through actionable insights.
• Developed scalable dimensional models and dashboards for multi dimensional analysis of campaign performance, enabling efficient tracking across time, channel, and customer segments.

05/2021 - 05/2022
Data Engineer
Wefox Germany GmbH (Banken und Finanzdienstleistungen, 250-500 Mitarbeiter)

Worked on connecting various data sources, creating ETL jobs, and building data warehousing solutions using technologies like BigQuery, Snowflake, and Airflow.
Connect Google Analytics and Firebase with BigQuery and export all data into BigQuery.
• Create python jobs get data from BigQuery large tables and load data into s3 storage in parquet format.
• Create jobs in Snowflake to load data into landing table in json string then parse data into tabular format and load into the internal table of warehouse
• Build data lake and data plateforms for stakeholders.
• Create Data warehousing and data modeling in Snowflake.
• Create dynamic Airflow dags so that we only need to add information in config file for new dags.
• Responsible to monitoring the alerts of applications using Grafana dashboard and respond on time.
• Create dashboard and reports in Apache superset and implement user level data governance.
• Create Python Glue crawlers for partitions files.
• Designed and implemented robust ETL pipelines in Apache NiFi to streamline data processing workflows.
• Developed custom Python scripts for advanced data parsing, enhancing data transformation capabilities within NiFi.

08/2018 - 04/2021
Big Data Consultant
Systems Limited

Focused on migrating ETL jobs to Spark, optimizing data processing pipelines, and implementing AWS architectures for automated infrastructure deployment.

09/2017 - 08/2018
Data Engineer
FutureNow Technology

Handled large-scale data processing tasks, developed NIFI pipelines for SQL to S3 data flow, and worked on optimizing Spark applications for improved performance.

Reisebereitschaft

Weltweit verfügbar
Profilbild von HafizRanaSherHassanAli Ibrahim Experienced Data Engineer with Big Data Expertise aus Berlin Experienced Data Engineer with Big Data Expertise
Registrieren