Data Engineer - Spark & Hadoop

München, Bayern  ‐ Vor Ort
Dieses Projekt ist archiviert und leider nicht (mehr) aktiv.
Sie finden vakante Projekte hier in unserer Projektbörse.

Beschreibung

Glocomms is urgently searching for a Data Engineer to join a Global Enterprise business based in Munich.

Key Responsibilities
•As (Senior) Big Data Engineer / Developer you will be closely working with IT architects to elicit requirements, to optimize the system performance as well as to advance its technological foundation.
•Manage very large-scale, multi-tenant and secure, highly-available Hadoop infrastructure supporting rapid data growth for a wide spectrum of innovative internal customers
•Provide architectural guidance, planning, estimating cluster capacity, and creating roadmaps for Hadoop cluster deployment.
•Install Hadoop distributions, updates, patches, version upgrades
•Design, implement and maintain enterprise-level security (Kerberos, LDAP/AD, Sentry, etc.)
•Develop business relevant applications in Spark, Spark Streaming, Kafka using functional programming methods in Scala.
•Implement statistical methods and machine learning algorithms to be executed in Spark applications, which are automatically scheduled and running on top of the Big Data platform
•Identify new components, functions and features and drive from exploration to implementation
•Create run books for troubleshooting, cluster recovery and routine cluster maintenance
•Troubleshoot Hadoop-related applications, components and infrastructure issues at large scale.
•Design, configure and manage the strategy and execution for backup and disaster recovery of big data.
•3rd-Level-Support (DevOps) for business-critical applications and use cases
•Evaluate and propose new tools and technologies to meet the needs of the global organization.
•Work closely with infrastructure, network, database, application, business intelligence and data science units.

Key Requirements, Skills, and Experience
•University degree in computer science, mathematics, business informatics or in another technical field of study.
•Deep expertise in distributed computing and the factors determining and affecting distributed system performance
•Experience with implementing Hadoop clusters in a large scale environment, preferably including multi-tenancy and security with Kerberos.
•Excellent hands-on working experience with Hadoop ecosystem for at least 2 years, including Apache Spark, Spark Streaming, Kafka, Zookeeper, Job Tracker, HDFS, MapReduce, Impala, Hive, Oozie, Flume, Sentry, but also with Oracle, MySQL, PSQL
•Strong expertise in functional programming, object-oriented programming and scripting, i.e. in Scala, Java, Ruby, Groovy, Python, R
•Proficiency with IDEs (IntelliJ IDEA, Eclipse, etc.), build automation (Maven, etc.) and continuous integration tools (Jenkins, etc.)
•Strong Linux skills; hands-on experience with enterprise-level Linux deployments as well as shell scripting (bash, tcsh, zsh)
•Well versed in installing, upgrading & managing distributions of Hadoop (CDH5x), Cloudera Manager, MapR, etc.

20 Hours Per week.
6 Month Contract
Rate: Negotiable

Please send your CV to if you are interested in this position.
Start
08.2018
Dauer
6 Monate
(Verlängerung möglich)
Von
Glocomms
Eingestellt
18.07.2018
Ansprechpartner:
Karmen Ledgister
Projekt-ID:
1600554
Vertragsart
Freiberuflich
Um sich auf dieses Projekt zu bewerben müssen Sie sich einloggen.
Registrieren