Junior Data Lake Infrastructure Dev (19097)

5/2020 (6m+)
Kontrakt přes CP
120 000 CZK
« zpět
Tato pozice není aktuálně dostupná
Are you looking for a job that has a worldwide reach and deep purpose? Then come join a global healthcare leader as a Lake Infrastructure (Big Data) Developer.

Location: Prague
Form of cooperation: Freelance contract
Start of cooperation: 01/2020

About the project: 
Create awesome digital products and enjoy a reward that technology careers don’t often bring: the satisfaction of helping to save lives. As a part of Big Data Platform team you will contribute to the client‘s Data Lake and cooperate with other teams to tackle the biggest opportunities at the intersection of healthcare, information and technology. 

What will you be working on: 
- be in charge of development, deployment and maintenance of Data Lake infrastructure
- apply best practices, provide infrastructure as-a-code, automate, execute releases, test and monitor
- responsible for implementation and ongoing administration of Data Lake infrastructure
- design, develop, test and implement a migration of Hadoop services to AWS EMR, Glue and Redshift services over S3
- performance tuning of Data Lake infrastructure
- automation of manual tasks using Ansible
. collaborating with application teams to install operating system and service updates, and patches
- Hadoop services support and maintenance - HDFS, Hive, HBase, Spark and Kafka
- Elastic Cloud Enterprise platform deployment, monitoring and support
- TICK Stack platform deployment, monitoring and support
- research and recommend technical and operational improvements for improved reliability and efficiencies
- adopt new technologies and maintain interest in automation and continuous delivery excellence

What you need to know: 
- familiarity with UNIX/LINUX system administration
- knowledge: Bash and Python
- basics of networking and infrastructure
- understanding of memory management
- basic architecture understanding of distributed systems such as Hadoop
- troubleshooting skills
- willing to learn and adapt
- ability to communicate in English

Nice to have:
- knowledge of Hadoop eco system such as YARN, MapReduce, HDFS, HBase, Zookeeper and Hive
- experience in administering (No)SQL database
- familiarity with opensource configuration management and deployment tools such as Puppet, Chef or Ansible
- knowledge of Java applications troubleshooting
- knowledge of agile development
- knowledge of ElasticStackor TICK Stack
- knowledge of AWS EC2, EMR, Glue, Redshift, S3 services