Job Search
Java Big-data Engineer and Administrator (CST timezone) IRC151471
Job: | IRC151471 |
Location: | Poland - Krakow |
Designation: | Senior Software Engineer |
Experience: | 3-5 years |
Function: | Engineering |
Skills: | Big Data, Hadoop, Hive, Kafka, Spark |
Remote | Yes |
Description:
Our client delivers personalized experiences, cross-channel messaging and loyalty programs that add value to the customer relationships. The the product enables marketers to rapidly scale and meet the demands of the modern consumer. Marketers can easily deploy targeted and triggered cross-channel campaigns, compile data from multiple sources, and strategically promote desired customer behaviors – effectively increasing revenue and enhancing customer experiences.
Requirements:
- 5+ years of Java Development experience or 3+ years in a Big-data cluster Administration role
-
3+ years in a Big-data Engineer role;
-
Sound knowledge of Hadoop Architecture and distributed data environments;
-
Excellent functional knowledge and experience with big data components such as Kafka, Spark,Hadoop, Hbase, Hive;
-
Experience in optimizing hadoop/hive settings for maximum throughput (includes compaction; hive metadata; TeZ Containerization;);
-
Experience in shell scripting, have knowledge of scripting languages such as Python and be exposed to programming languages such as Java;
-
Knowledge of Table definitions, file formats, UDF, Data Layout (Partitions and Buckets);
-
Monitor Cluster performance, resource usage, backup and mirroring, provisioning and automation (Administration);
-
ProdOps experience on Storm/Kafka.
-
Experience in analyzing MapReduce jobs logs and error handling;
-
AWS/Cloud: Sound knowledge of Cloud basics: S3, EMR. Need experience in identifying opportunities to improve processes, operations and automation using IaC;
-
Experience ingesting data from multiple data sources such as REST API, SFTP flat files, Streaming data using Apache Storm, Kafka Streams or others etc.
-
Degree in Computer Science, IT, or similar field; a Master’s is a plus;
-
Data engineering certification (e.g AWS Certified Big Data – Specialty) is a plus;
-
Experience with Agile methodology;
-
English: Upper-intermediate, both verbal and written.
Will be a plus:
Preferences:
Python, Java, AWS, Hbase, Storm
Job Responsibilities:
-
Participate in developing high load MapReduce platform hosted on-premise and AWS.
-
Participate in architectural decision making on on-premise migration to AWS EMR
-
Be ready to assist in integration of new data streaming sources (like REST API, SFTP flat files)
-
Direct communication with the customer
What We Offer
Exciting Projects: With clients across all industries and sectors, we offer an opportunity to work on market-defining products using the latest technologies.
Collaborative Environment: You can expand your skills by collaborating with a diverse team of highly talented people in an open, laidback environment — or even abroad in one of our global centers or client facilities!
Work-Life Balance: GlobalLogic prioritizes work-life balance, which is why we offer flexible work schedules.
Professional Development: We develop paths suited to your individual talents through international knowledge exchanges and professional certification opportunities.
Excellent Benefits: We provide our employees with private medical care, sports facilities cards, group life insurance, travel insurance, relocation package, food subsidies and cultural activities.
Fun Perks: We want you to feel comfortable in your work, which is why we create good working environment with relax zones, host social and teambuilding activities and stock our kitchen with delicious teas and coffees!