Data Engineer (Scala)

Job description

Zooplus AG is Europe’s leading online retailer of pet supplies. With over 6,8 million active customers in more than 30 European countries and 1,342 billion Euro revenue in 2018, zooplus is comfortably the market leader in the online segment. At zooplus, we believe e-commerce to be the sales model of today and for the future. E-commerce has become one of the fastest developing areas of business. Modern logistics centers, and the use of big data and personalization, mean that online sales can be handled with increasing efficiency. 


We’re an agile company, our IT Culture fosters cross-functional teams (a.k.a. Feature Temas) where the focus is on product specialization, and trust is one of our core values. 


Summary:

Zooplus is building an in-house big data platform that will allow us to scale our data strategy for the future. In this role you will have the opportunity to be working in the development of a cutting-edge data platform that will be core for the company. You will integrate into an experienced agile team (using Scrum plus other agile frameworks) that is building our data platform from the ground up on the AWS cloud. You will also develop the integration of different data sources covering the whole ecosystem of Big Data technologies. 


Our tech stack includes (but is far from limited to):

  • Apache Spark with Scala 
  • AWS EMR, Glue, S3, Aurora, Lamdba, Athena, DMS 

  • Parquet, Avro 

This is an almost greenfield project where the team holds full ownership, from idea to production: you build it you run it.


What you will be responsible for:


  • Working on the software development of our Data Platform, closely together in team of Data Architects, Data Engineers and Data Scientists in an agile environment. 

  • Develop the core components of the Data Platform with focus on good practices, testability and reliability. 

  • Create and test ETLs and tools to automate ETLs for the data migration into the platform 

  • Create the endpoints that will leverage structured and unstructured data integration from multiple data sources into the data platform. 

  • Work with other cross-functional data and product teams and build efficient and scalable data solutions 

Requirements

What we are looking for:

To work together in an optimal way in this position, we wish your qualifications to include

  • BS/M. Tech in Computer Science or related field
  • Highly proficient in Apache Spark with Scala
  • Solid software development skills
  • Expertise in data engineering and DWH/ETL best practices (experience building a Data Lake is a plus)
  • Experience working with data technologies in Amazon Web Services (EMR, Kinesis, S3, Aurora, Glue, Lamdba, Athena).
  • Comfortable with the idea of cross-functional teams
  • Fluent English 

What we are offering you

Become a part of our success story and seize the opportunity to take on a real challenge in a dynamically growing company where there is huge scope for development and short decision-making processes. We are offering you a versatile, international-facing role in our motivated team with colleagues from all over Europe. Our Spanish office is at a great location in Madrid, with great infrastructure links. Additionally we offer competitive benefits such as:


  • Competitive salary
  • 28 days’ vacation (plus Dec 24 and 31 when they fall on a labor day)
  • Medical insurance with a dental plan
  • Flexible working hours
  • Free drinks and fresh fruit
  • Discount in zooplus shop 


Did we make you curious?

Then send us your application - resume, motivation letter, indication of your desired salary and possible starting date in Spanish or English via our online application form.