Requirements
- 3+ years data engineering experience (level dependent) with real pipeline delivery beyond ad-hoc scripts.
- Strong Python + SQL; comfortable building transformations, validation tooling, and pipeline glue code.
- Practical streaming/CDC fundamentals (ordering, duplication, replay, idempotency) and Kafka ecosystem experience.
- Familiar with lakehouse/storage and query layers (e.g., Hudi/Iceberg/Delta, Trino/Hive/Postgres) and how to make datasets usable.
- Comfortable working in Kubernetes/container environments and documenting decisions clearly.
- Eligible to work in Germany; EU/NATO citizenship preferred and export-control screening applies.
Nice to Have
- Great Expectations or similar data quality tooling; metadata/lineage platforms (OpenMetadata/DataHub/Atlas).
- Experience shipping in on-prem or air-gapped environments; governance/policy awareness for regulated customers.
- German language (B1+) and/or experience with OSINT/GEOINT/multi-INT data shapes.
Additional Information
- Eligible to work in Germany
- EU/NATO citizenship preferred
- Export-control screening applies
- Equipment budget provided
- Learning budget provided
- Regular in-person sprints in Berlin