Profilbild von Alexandre ManhesSavio Senior Solution Architect/Software and Data Engineer - DevOps, Big Data, Cloud aus Muenchen

Alexandre Manhães Savio

verfügbar

Letztes Update: 15.04.2024

Senior Solution Architect/Software and Data Engineer - DevOps, Big Data, Cloud

Abschluss: PhD in Computer Science
Stunden-/Tagessatz: anzeigen
Sprachkenntnisse: deutsch (Grundkenntnisse) | englisch (verhandlungssicher) | portugiesisch (Muttersprache) | spanisch (Muttersprache)

Dateianlagen

cv-alexandre-savio-cs-en-long_150424.pdf

Skills

  • Experiences
    • 2023-2024: MLOps Engineer, Bayer. Build tools and services for the AI development team to run ML models on production for different precision farming applications. Based on Terraform, Serverless, and Python.
    • 2023: Senior Data Engineer, KONUX. Build a real time data pipeline based on PySpark to process UK Railways network data to create a network digital twin.
    • 2022: Senior Data Solution Architect, Klarna. Build a real-time Slack alerting system based on Rockset, Kafka, and Kubernetes using AI. A tool to monitor the Klarna service integrations from their customers.
    • 2021: Senior Data Solution Architect, Personio. Build serverless data pipelines in the cloud to aggregate data from different third-party data providers for business reporting.
    • 2021: Data Engineer, CARIAD, Remote. As a data engineer I improved the reliability and testability of the data pipelines running in Airflow for a Points of Interest data aggregation system for the Volkswagen ID.3 and ID.4.
    • 2019-2021: Cloud Applications/Data Engineering Team Lead, Nordcloud GmbH, Munich, Germany. As team lead I manage projects staffing, internal management, development processes, and knowledge sharing. Pre-sales and writing cloud solution project proposals. Implement integration and performance tests for a REST API in Go. Amazon Elasticsearch automated data ETL for reporting and data analytics. Cloud infrastructure with AWS CDK: CI/CD pipeline, step functions for data processing, monitoring, and alerting. Azure IoT PoC and Azure AKS tu run Dask and JupyterHub at scale to run datascience jobs.
    • 2018-2019: Cloud Applications Developer, Nordcloud GmbH, Munich, Germany. Technical lead in a project in the automotive industry to build a fast big data ingestion system for factory production data in Python based on AWS Step Functions to automate the data ingestion for a data warehouse in AWS Redshift and a data lake with AWS Glue. Infrastructure defined with Terraform, CI/CD pipeline with code linting, static type checking, and integration tests. Serverless Framework for the lambda functions, step functions, triggers, and alerts. Automated monitoring and alerting with AWS Cloudwatch and SNS.
    • 2017-2018: Cloud Application Engineer, Software Defined Inverter, Siemens Corporate Technology, Munich, Germany. Using Python and Terraform for IoT project: RESTful API test driven development, IoT tooling for production, and cloud infrastructure in AWS.
    • 2017: IT Architect, MaibornWolff Gmbh, Munich, Germany. BMW Log-Analyzer: software engineering and DevOps for IT infrastructure with Ansible, Python, Kafka, and ELK stack for log data stashing and analysis. BMW Augmented Reality Tracking: C++ application to measure and estimate in real time IMU data and UR robotic arm to simulate human head movement for car driver HUD.
    • 2015-2017: Post-doctoral researcher, Klinikum rechts der Isar, TUM, Munich, Germany.  Brain PET-MRI processing for computer-aided diagnosis systems. Specially using resting-state fMRI, DTI and FDG-PET. Brain image data pre-processing and analytics pipeline done with Python and Matlab. Medical image processing and time-series analyses for age-related dematiae detection and understanding. AWS for big data processing. Scientific publications and lecturing in intensive workshops for scientific staff focused on open source software, bash, git, and programming with Python. Using Scikit-learn and pipelines with nipype.
    • 2014-2018: Board member and organizer of EuroPython 2015 and 2016 in Bilbao, Spain. Software and Data Carpentry Instructor:
      Intensive courses for scientific staff focused on open source software, bash, git, and programming
      with Python. 2016 in Bilbao, Spain. Software and Data Carpentry Instructor.
    • 2015: Post-doctoral researcher, Wrocław University of Technology, Wrocław, Poland. Machine learning applied to brain and abdominal MRI for computer-aided diagnosis systems. Specially using resting-state fMRI and time-series analysis. International research collaborations and publications.
    • 2013-2015: Post-doctoral researcher, Computational Intelligence Group (UPV/EHU), San Sebastián, Spain. Machine learning approaches applied to resting-state fMRI connectivity using Python. Setup and provisiong of an HPC cluster for the department's brain image processing cluster. Using Scikit-learn and pipelines with nipype.
    • 2008-2013: PhD Student, Computational Intelligence Group (UPV/EHU), San Sebastián, Spain. Multimodal brain MRI processing. Machine learning for feature selection, dimensional reduction, and classification. Bayesian sampling methods. Lecturing master degree. Using Scikit-learn and pipelines with nipype.
    • 2007-2008: Software Engineer, InnovaeVision S.L., San Sebastián, Spain. Java and Hibernate web service with frontend application for certification company. Embedded GNU/Linux C++ application for specific hardware for medical treatment.
    • 2004-2007: Software Engineer, JADE Systems S.L., Pamplona, Spain. Systems and network administrator. Windows C++ programming for aeronautical material calculation software for wind energy turbine structures.
  • Studies
    • PhD in Information Engineering, Computational Intelligence Group (GIC), UPV/EHU, San Sebastián, Spain. Approved with honors (Apto cum laude). Experiments of Machine Learning for Neurodiagnosis, mentored by Prof. Dr. Manuel Graña Romay.
    • BSc. and MSc. in Information Engineering, University of the Basque Country (UPV/EHU), San Sebastián, Spain.
      Graduation project: "Experiment with Hidden Markov Models for the Spanish sign-language recognition". Approved with honors.
  • Courses, awards and others
    • AWS Cloud Developer Associate certificate. Valid until end of 2020.
    • Txekintek/Ekintzaile, BIC Berrilan, San Sebastián, Spain. Public funding for medical image processing startup.
    • XXIV Premio Toribio Echeverria Award and XIII Manuel Laborde Werlinden Award, Basque Country Government, Eibar, Spain. 2 awards for a startup idea and project.
    • 6 months internship at FMRIB, The University of Oxford, UK. On modeling of brain diffusion data with Bayesian approaches and model selection processes.
    • Community work. Co-organizer of EuroPython 2015, 2016, and 2018, EuroSciPy 2019, and Remote Python Pizza. Speaker
      at EuroPython 2016, PyData Florence 2017, and PyconWeb 2018.

Projekthistorie

04/2023 - bis jetzt
Software Engineer
Bayer AG (Pharma und Medizintechnik, 1000-5000 Mitarbeiter)

Python, AWS, and MLOps

10/2022 - 02/2023
Data Engineer
KONUX GmbH (Transport und Logistik, 50-250 Mitarbeiter)

PySpark and AWS data engineering in the R&D team.

11/2021 - 07/2022
Software Engineer
Klarna Bank AB (Banken und Finanzdienstleistungen, 1000-5000 Mitarbeiter)

Real-time anomaly detection system using Kafka, AWS Kinesis Data Analytics, running services in Kubernetes, and storing the hot timeseries data in Rockset.

I was in collaboration with the team on designing the system. I implemented the CI/CD pipelines in Jenkins, I wrote an async Rockset client and the async systems to run the detection processes and APIs, and also sending alerts through Slack. I collaborated on the Docker images configuration and deployments, and implementing the Kinesis Data Analytics process.

10/2021 - 12/2021
Data Engineer
Personio GmBH (Internet und Informationstechnologie, 500-1000 Mitarbeiter)

Build ETL pipelines on AWS Step Functions using Terraform and Python. Improve automation of Continuous Integration and Deployment. Implement and coordinate the migration of the development environment to a production environment. Migrate Athena tables. Increase code quality and test coverage. Implement an API for ingestion of Google Spreadsheets into Athena.

07/2021 - 11/2021
Software and Data Engineer for POI ingestion and management service
Volkswagen AG (Automobil und Fahrzeugbau, >10.000 Mitarbeiter)

Build new features, improve code quality, maintainability, and test coverage of Airflow pipelines to manage Points-of-Interest (POI) information for electric vehicles.
Build a street parsing module based on trained natural language processing machine learning models.

01/2019 - 02/2021
Data Warehouse
Nordcloud Deutschland GmbH (Internet und Informationstechnologie, 500-1000 Mitarbeiter)

Design, implementation, and testing of a data ingestion (ETL) pipeline, to move relational data from
AWS S3 into AWS Redshift, then back to AWS S3 for datalakes. In a team of 2 developers and a
product owner, we built using Python, SQLAlchemy, and Boto3 on top of AWS Step Functions,
AWS Lambda, and AWS ECS, pipelines for data ingestion that ran every 10 minutes to ingest
~500MB of diverse relational data per run for 35 production factories around the globe.
Other parts of the solution was also running data transformation, and finally moving the data
back to S3 in parquet format to create a data lake available to end users through Athena and
Microstrategy.

For some maintenance we were using AWS EMR to process data, clean up parquet files, fixing
data partitions, and moving data to other regions or buckets.

11/2017 - 11/2018
Junelight project at Siemens CT
Siemens AG (Internet und Informationstechnologie, >10.000 Mitarbeiter)

Design, development, and testing of a REST API for a web and mobile app for the fleet management of an IoT product from Siemens. This API would use best practices using automated code checks and testing, automated database migration, and clean REST API design. Built to run at scale in AWS with API Gateway, CI/CD pipelines with integration and end-to-end tests. It was developed in Python using the Falcon API framework and SQLAlchemy as an ORM against PostgreSQL. Lambda functions and EC2 for monitoring and data pipelines. I was also in charge of factory tooling systems, which was a small web UI running on a Raspberry Pi used to install firmware to a fresh device, test it, register it on the systems through the API, and print a QR code for device identification.

04/2017 - 09/2017
Service logs storage modernization for a big car manufacturer
MaibornWolff GmbH (Internet und Informationstechnologie, 250-500 Mitarbeiter)

This was a project for a big car manufacturing company where we automated the provisioning of 3
on-premise clusters with Kafka and the ELK stack in 3 different regions.
It was a team of 4 people, so I was very involved in a wide range of tasks, including the
logging library maintenance, the Ansible project for automating the cluster provisioning, and the monitoring
solution. For Kafka we implemented monitoring channels for the systems, and I also
maintained a logging library in Java based on Log4J to support other developers to ship their service
logs through Kafka.

Reisebereitschaft

In der Stadt München mit einem Radius von 50 km verfügbar
I am flexible to travel once in a while.

Sonstige Angaben

Before calling for a position, please write to me here. I am usually busy and do not pick up phone calls without a conversation here first.
Profilbild von Alexandre ManhesSavio Senior Solution Architect/Software and Data Engineer - DevOps, Big Data, Cloud aus Muenchen Senior Solution Architect/Software and Data Engineer - DevOps, Big Data, Cloud
Registrieren