DHL Information Services (Europe) s.r.o.
JOB DESCRIPTION About DHL IT Services and the AI & Analytics team: At DHL IT Services, we are designing, building and running IT solutions for the whole DPDHL globally.The AI & Analytics team builds and runs solutions to get much more value out of our data. We help our business colleagues all over the world with machine learning algorithms, predictive models and visualizations. We manage more than 46 AI & Big Data Applications, 3. active users, 87 countries and up to ,, daily transaction.Integration of AI & Big Data into business processes to compete in a data driven world needs state of the art technology. Our infrastructure, hosted on-prem and in the cloud (Azure and GCP), includes MapR, Airflow, Spark, Kafka, jupyter, Kubeflow, Jenkins, GitHub, Tableau, Power BI, Synapse (Analytics), Databricks and further interesting tools.We like to do everything in an Agile/DevOps way. No more throwing the “problem code” to support, no silos. Our teams are completely product oriented, having end to end responsibility for the success of our product. Who do we look for: Currently, we are looking for Senior Data Engineer. In this role, you will have the opportunity to design and develop solutions, contribute to roadmaps of Big Data architectures and provide mentorship and feedback to more junior team members. We are looking for someone to help us manage the petabytes of data we have and turn them into value.Does that sound a bit like you? Let’s talk! Even if you don’t tick all the boxes below, we’d love to hear from you; our new department is rapidly growing and we’re looking for many people with the can-do mindset to join us on our digitalization journey. Thank you for considering DHL as the next step in your career – we do believe we can make a difference together! What will you need: • University Degree in Computer Science, Information Systems, Business Administration, or related field • 3+ years of experience in the Data Engineer role • Strong analytic skills related to working with structured, semi structured and unstructured datasets • Hands-on experience with Data Lake/Big Data Projects implementation on On-premise or Cloud platforms (preferably Azure) • Hands-on experience with Docker and Kubernetes and related ecosystem tooling on-prem and in public clouds • Hands-on experience with public clouds (GCP, Azure, Google), with specific focus on using them for Data Lakes • Experience working with Big Data Technologies – e.g. Spark, Kafka, HDFS, Hive, Hadoop distributions (Cloudera or MapR) • Experience with streaming platforms/frameworks such as Kafka, Spark-Streaming, Flink • Good Programming skills (Java/Scala/Python) • Advanced working SQL knowledge and experience with relational databases, query authoring (SQL), as well as working familiarity with a variety of databases • Proven experience in building and optimizing big data pipelines, architectures and data sets • Experience in performing root cause analysis on internal and external data and processes to answer specific business questions and identify opportunities for improvement • Experience in building processes supporting data transformation, data structures, metadata, dependency and workload management • A successful history of manipulating, processing and extracting value from large disconnected datasets • Experience with Git, CI/CD and Good to have Containerization like Docker or Open shift You should have: • Certifications in some of the core technologies • Ability to collaborate across different teams/geographies/stakeholders/levels of seniority • Customer focus with an eye on continuous improvement; • Energetic, enthusiastic and results-oriented personality • Ability to coach other team members, you must be a team player! • Strong will to overcome the complexities involved in developing and supporting data pipelines Language requirements: • English – Fluent spoken and written (C1 level) What we offer: • Great team of IT professionals and possibility of technical development • Modern offices in Chodov • Home office possibilities • Permanent contract • CAFETERIA employee benefit program with wide selection of benefits from Edenred • Extra week of holiday (25 days/year), 6 Self-sickness days/year, Full salary compensation for up to 10 days absence due to illness per calendar year, Lunch vouchers fully covered by company • Multisport card, mobile and laptop, fruit days, sport clubs for employees, Referral program……For more details feel free to contact .