Zum Inhalt springen

(Senior) Data Engineer – AI Booster Team (f|m|d) - Bewerbungsfrist abgelaufen 21.11.

  • On-site
    • Berlin, Berlin, Germany
  • Technology & Data

Job description

At idealo, Generative AI (GenAI) is becoming a multiplier across every team. The AI Booster Team is our internal technical competence center: we pair with product teams, build reusable GenAI building blocks and share best practices company-wide.

As a Data Engineer in our AI Booster Team, you’ll be at the center of this transformation: designing data pipelines, integrations, and automations that power GenAI use cases at scale. You’ll combine classic data engineering skills with modern AI infrastructure, ensuring that product teams have instant, reliable access to “LLM-ready” data and the automation tools they need to move fast.

This position is available full-time or part-time.

About your new role

  • Build robust pipelines: Ingest, transform, and unify data from APIs, databases, files, and streams into analytics- and LLM-ready formats. 

  • Engineer integrations: Develop connectors and orchestrations (Airflow, n8n, Step Functions) that move data securely and efficiently between warehouses, APIs, CMSs, and GenAI services. 

  • Operate modern data stores: Manage vector databases and feature stores to enable fast, reliable retrieval for RAG and ML use cases. 

  • Ensure reliability & cost-efficiency: Implement data quality checks, lineage tracking, monitoring, and FinOps practices. 

  • Enable self-service: Provide reusable workflows, templates, and automation components that empower product teams to build on top of your data infrastructure. 

  • Coach & collaborate: Share best practices, write playbooks, and guide teams in building scalable data-driven and GenAI-enabled workflows. 

Skills & Requirements

  • 3+ years in data engineering or MLOps, delivering production-grade data integrations. 

  • Strong experience unifying heterogeneous data sources (SQL/NoSQL, APIs, streams). 

  • Advanced Python & SQL skills; comfortable with Spark/Glue, Kafka/Kinesis, schema evolution. 

  • Hands-on with AWS services (S3, Glue, Redshift, Lambda, SageMaker, Bedrock, CDK/Terraform). 

  • Familiarity with vector stores and embedding pipelines for RAG. 

  • Strong focus on observability, reliability, and cost control. 

  • Excellent communication skills for enabling and coaching non-data specialists. 

  • We’re keen to see evidence of exceptional achievement - perhaps you’ve scaled a personal project to thousands of users, published influential research, ranked highly in competitive arenas (e.g. sports, Kaggle, hackathons) or maintain widely-used open-source libraries. Tell us what makes you stand out! 

You don’t tick every single box? No worries! We hire people, not checklists, and value motivation to grow.

Job requirements

#LI-AJ

On-site
  • Berlin, Berlin, Germany
Technology & Data

or

Apply with Linkedin unavailable
Apply with Indeed unavailable

About idealo

idealo is one of Europe's leading online comparison platforms for products. With more than 2.5 million page views per day, over 600 million product offers from around 50,000 merchants for the comparison platform, we are one of the largest e-commerce websites in Germany! In 2000, we started out with the mission of helping consumers make the best purchasing decisions. Right in the heart of Berlin, about 700 employees from almost 60 nations are supporting our users to find the best deal. idealo is majority-owned by the transatlantic media company Axel Springer SE.

Information for applicants with severe disabilities:

As an applicant with a severe disability, you can get advice from our Representative Body for Severely Disabled People (SBV) during the application process. Please inform us regarding this topic. For the inclusion of people with disabilities, our office is accessible without steps. When you start with us, we will see what else is needed!