
Since 2017, the number of Python users has been increasing by millions annually. The vast majority of these people leverage Python as a tool to solve problems at work. Our mission is to make them autonomous when they create and use data in their organizations. For this end, we are building an open source Python library called data load tool (dlt). Our users use dlt in their Python scripts to turn messy, unstructured data into regularly updated datasets. It empowers them to create highly scalable, easy to maintain, straightforward to deploy data pipelines without having to wait for help from a data engineer. We are dedicated to keeping dlt an open source project surrounded by a vibrant, engaged community. To make this sustainable, dltHub stewards dlt while also offering additional software and services that generate revenue (similar to what GitHub does with Git). dltHub is based in Berlin and New York City. It was founded by data and machine learning veterans. We are backed by Dig Ventures and many technical founders from companies such as Hugging Face, Instana, Matillion, Miro, and Rasa.

Since 2017, the number of Python users has been increasing by millions annually. The vast majority of these people leverage Python as a tool to solve problems at work. Our mission is to make them autonomous when they create and use data in their organizations. For this end, we are building an open source Python library called data load tool (dlt). Our users use dlt in their Python scripts to turn messy, unstructured data into regularly updated datasets. It empowers them to create highly scalable, easy to maintain, straightforward to deploy data pipelines without having to wait for help from a data engineer. We are dedicated to keeping dlt an open source project surrounded by a vibrant, engaged community. To make this sustainable, dltHub stewards dlt while also offering additional software and services that generate revenue (similar to what GitHub does with Git). dltHub is based in Berlin and New York City. It was founded by data and machine learning veterans. We are backed by Dig Ventures and many technical founders from companies such as Hugging Face, Instana, Matillion, Miro, and Rasa.
Product: Open-source Python library 'dlt' for building EL/ELT pipelines; platform product 'dltHub' planned
Headquarters: Berlin and New York City
Founders: Four co-founders including Matthaus Krzykowski (CEO) and Marcin Rudolf (CTO)
Funding: $1.5M pre-seed (Jul 2023), lead investor Dig Ventures
Team size: 38 employees (stated)
Data engineering / EL/ELT for Python-first teams; pipeline automation, data ingestion, and governance.
Data infrastructure / Data engineering
$1.5M
Round announced July 20, 2023
“Backed by Dig Ventures and other venture/backer mentions including Foundation Capital and technical founders”
| Company |
|---|
About dlthub
dltHub is built on dlt, the most widely adopted open-source Python library for data loading with ~4 Million monthly downloads. We are building the next generation of AI-native data tooling on top of our strong OSS foundation.
We move and iterate fast, and ship with a strong bias toward experimentation, automation and user-centric AI workflows. We’re a highly dynamic, fast-moving team founded by experienced data and ML engineers, with roots in Berlin and a growing presence in the US.
Founded by experienced data and ML engineers, we're backed by , , , and technical founders from , , , , , , , and . Our team is driven by clear values: we speak with courage, build what matters, automate relentlessly, show up with energy, deliver on commitments, and win together.
We are building the next generation of AI-native data tooling - building our AI-first data platform on top of our popular open-source data ingestion core.
We are looking for an entrepreneurial software engineer based in Berlin who thrives in a fast-paced environment with a high degree of autonomy and ownership. You will work on solving data platform challenges for our customers: ingestion, storage, performance, reliability, observability and AI-assisted developer experience. If you enjoy working with an ambitious and driven team and care deeply about code quality, data systems, and building tools that engineers enjoy using, this role is for you.
🗂️ Key Responsibilities
What We’re Looking For
✅ Qualifications
7+ years of experience in software development and/or data engineering.
Experience in designing scalable systems.
Strong Python skills: clean design, typing, testing, and maintainable code.
Solid understanding of data engineering fundamentals: data ware-/lakehouses, object storage, data ingestion and transformation patterns.
Good SQL knowledge and understanding of databases, transactions, and concurrency.
Experience working with modern Python data tools (e.g. DuckDB, Arrow, Iceberg, delta-rs, vector stores, storage buckets).
Clear written and verbal communication skills.
Team-first mindset and high degree of autonomy and ownership.
Passion and curiosity for AI-assisted software development workflows
Based in Berlin and open to working onsite regularly (3-4 days per week).
🎯 Nice to Have
In our work culture, we value each other’s autonomy and efficiency. We have set hours for communication and deep work. We like automation, so we automate our work before we automate the work of others.