Beschreibung
Aufgaben:Developing data-pipelines for RDBMS and NoSQL environments
Establishing data integration from on-prem to the cloud
Implementing end-to-end cloud native monitoring and alerting frameworks
Making use of serverless computing solutions (e.g. App Engine, Cloud Functions)
Usage of advanced scheduling mechanisms (e.g. Apache Airflow)
Enhancement of CI/CD pipelines
Implementing the microservice architecture pattern for data solutions
Requirements engineering together with our business product teams up to technical concepts for bringing these concepts live
Business customer consulting for modelling data structures / data marts / reporting solutions
End-to-end responsibility from the creation of the technical concept to the implementation
Skills:
Practical experience with data warehouse / data lake systems
Proficiency in object oriented programming languages like Java or Ruby and knowledge in data-science programming languages like Python or R
Experience with frameworks and packages like Pandas, Matplotlib, Scikit-learn, TensorFlow or similar is desirable
Proficiency in SQL
First experiences with cloud-technologies (preferable Google Cloud Platform or Microsoft Azure), NoSQL database technology and Apache Beam or Airflow
Start: asap , Ende: min. 6 Monate, Auslastung: 100% , Location: Remote