Data Platform Engineer

Vilnius

Vinted logo
Vinted
Apply now Apply later

Posted 1 week ago

Vinted is Europe’s biggest pre-loved fashion marketplace, with 37 million members in 13 countries. We want to make second-hand the first choice worldwide.
A team of over 700 people from Vinted’s offices in Vilnius, Berlin, Prague, and Utrecht develop and support our platform. We share a unique work culture that’s based on aiming high, taking ownership, co-creating, caring, and growth.
Vinted’s Data Platform Team makes sure everyone in our company has fresh data to make good decisions. At the moment there is one team working on the data platform. We are looking to scale this area to more teams and are hiring engineers to make that possible. You'll be working on improving the data platform to meet company needs: enable analysts to analyse ever-growing amounts of data (125+ TB of data at the moment, +16B new events per week) and help product teams with heavy data lifting to make our core product even greater. 
We are looking for someone who likes to solve problems related to data. As there are many unsolved problems in the domain, we are always on lookout for new techniques and technologies, we experiment a lot and use unconventional ways to solve problems. Knowledge on how database systems work is a big plus. Experience with Apache Spark is not mandatory, but very useful. We value pragmatism, big picture thinking, curiosity and problem solving skills. 
We expect you to be familiar with most, and have deep knowledge in at least a few of the following disciplines: Database Systems, Algorithms, Software Engineering, Systems Architecture, Big Data, Systems Scaling, Systems Performance Tuning, Computer Science. 
This is a mid / senior level position.

In this role you will:

  • Make sure that our batch and streaming data pipelines run smoothly and efficiently
  • Optimise data collection, storing and analysis on a large scale
  • Identify improvement opportunities in our data platform and implement them
  • Work closely with our staff and infrastructure engineers to research, experiment with and implement solutions based on new state of the art technology

  • Here are some of the technologies we use: Hadoop HDFS, Apache Spark (for ETL, Streaming, ML and ad-hoc analysis), Apache Kafka (message bus for tracking events and integration with our core product), Cloudera Impala (for run-time metric aggregation, quick ad-hoc querying). For most of the solutions we use Scala, Python and some tiny bit of Ruby.

About you:

  • Willing to learn and master new technologies and concepts
  • Experience building data platforms, preferably Hadoop based
  • Strong programming skills in a statically typed programming language, preferably JVM based (java, scala)
  • Scripting skills (bash, python or ruby)
  • Advantage: infrastructure knowledge and skills (Linux, Kubernetes, chef/puppet)
  • Advantage: data analysis skills

Perks:

  • Learning budget (13.2% from net salary)
  • Ability to work remotely up to 60% of the time25 working days of holidays
  • Remote workout classes
  • Mental- and emotional-health support from Mindletic
  • A canteen that serves homemade food at friendly prices
Depending on your experience, the salary range is €4900 - €7200 gross. If this sounds interesting - you may well be just who we need. Click 'apply for this job' and we will be in touch with you.
Job tags: Apache Bash Chef Hadoop Java Kafka Kubernetes Linux Puppet Python Ruby Scala Spark Streaming
Job region(s): Europe
Job stats:  1  0  0
  • Share this job via
  • or