DataChef

استخدام Data Engineer

DataChef
Amsterdam
امکان دورکاری

فرصت شغلی
درباره شرکت

تکنولوژی‌ها

    ScalaApache KafkaPythonJavaAmazon RDS

⚡️ The Shortest Possible Description

We are looking for an experienced, pragmatic data engineer to join our team and help in (probably) the largest implemention of an event-driven, serverless data mesh on AWS in Europe.

The mission is building a scalable data foundation for an Automative Marketplace 2.0; we are looking for A-players who are ready for the challenge and fun of creating such a unique foundation and stream processing backend using AWS services (Lambda, Glue, Lake formation,...), Spark (Scala/Python) and Kafka.

🙂 We like if you ... [a.k.a must have]

  • Have a real click with Our Core Values. If you’re nodding emphatically while reading it, you’ll probably fit right in, in which case, we can’t wait to hear from you. If your inner voice says "bla bla bla" while reading it, we want to save your and our time by not proceeding with interviews.
  • Are good consultants: have excellent communication skills to simplify and present concepts to other people. Show them how the future might be and help them participate in creating it. They don’t assume others know, so make everything (meetings, decisions, thoughts, code, etc) explicit & traceable.
  • Are doers, not talkers: we are a small team and our individual performance directly impacts the team outcome; so you need not only be taking initiatives but actually finishing what you’ve started. If you are not a level 4+ problem solver, don't even apply please.
  • Can demonstrate their solid technical skills in Python and/or Java (deep understanding of language internals, profiling, Testing methods). Knowing Scala is a plus.
  • Have 1+ years of hands-on experience with AWS services like lambda functions, EMR, Redshift, ElasticSearch, lake formation and Glue. Prior experience with a serverless framework like Chalice is a big plus.
  • Have a solid understanding (and preferably experience) of building pub-sub and asynchronous systems using Apache Kafka or any other messaging API like SQS, Kinesis, Celery, RabbitMQ, ActiveMQ.
  • Design & code defensively for the harsh real world and not for the happy path "Hello World" scenarios. They know missing, late and low-quality raw data is a fact and pipelines failures, replay/re-process are the norm, not a drama.
  • Can ingest new data sources (via REST APIs, file sharing etc) and maintain their ever-changing schemas.
  • Can analyze algorithms complexity and know data structures beyond "List and Stack" and pros/cons of using each for a problem.
  • Have been using Linux/Macos, Docker and Git in collaborative workflows for 2+ years.
  • Are fast movers: Our culture is "Go, go and go faster". Of course, you will break things by running fast which is understood and even appreciated. Just focus on learning fast and changing fast. And yes, we believe in agility and a distilled interpretation of agile manifesto.

💓 We love if you ... [a.k.a nice to have]

  • Have 5+ years of experience not only with developing greenfield projects from scratch but in an operational and live environment with strict high availability requirements.
  • Make quality a requirements issue: it is not enough to deliver something that works sometimes/maybe, we are building a mission-critical data platform. We love people who care about their craft and are proud about the quality of their code.
  • Write clean code that’s testable, maintainable, solves the right problem and does it well.
  • Know how to instrument their code for just-enough Logging, better monitoring and easier Debugging when it goes to production and operational environment.
  • Believe in DRY! They "Don’t Repeat Themselves". Have allergies for any kind of waste, manual repetitive not-automated task; well… after manually doing it a few times!
  • Understand CAP theorem and also know how to Design a resilient and partition tolerant service and its associated costs and tradeoffs.
  • Know about the latest developments in the Big Data community, combined with the ability to decide which of these are most relevant to our business and translate them into opportunities.
  • Contributed to open source: Make us happy with those green dots on GitHub!

👨‍🍳 This is us, bulletized.

  • DataChef: We're small, profitable, self-funded and growing company based in Amsterdam. If you believe that data can (and must) change the quality of life in companies (and humans who run those companies) then you will find yourself in your **tribe **here at DataChef.
  • We have a consultancy side, focused on delivering top quality projects on Big Data and machine learning and on AWS platform. Although, behind the scenes, we are working on a few SaaS ideas and products and aim at becoming a ۱۰۰٪ product company in next 3 years.
  • ۱۰۰٪ Open and Transparent company: Our role models are not giant corporates but relatively small yet happily successful companies like Basecamp, Buffer and ahrefs. We even share our salaries and benefits publicly - in case you're curious:
  • We thrive on technical excellence by hiring only the best and seeing ourselves at the beginning of the same success path as Databricks and Elasticsearch but only 10 years younger!
  • DataChef is an Equal Opportunity Employer – Minority / Female / Disability / Veteran / Gender Identity / Sexual Orientation.

مزایا

  • Remote working
  • 100% of Hardware & Software
  • MacBook Pro
  • Access to exclusive video and workshop training
  • Highest speed internet
  • Holiday & birthday gift
  • Coaching Sessions
  • Monthly Massage Subscription
  • Annual bonus