Kraken is hiring a Data Engineer to join our Data Platform team. You will build and maintain high-performance data pipelines and scalable data infrastructure to power Kraken's data ecosystem. Your work will ensure data reliability, quality, and accessibility for analytics, machine learning, and product innovation across our mission-focused company.
What You'll Do
- Build scalable and reliable data pipelines that collect, transform, load, and curate data from internal systems.
- Augment the data platform with data pipelines from external systems.
- Ensure high data quality for pipelines you build and make them auditable.
- Drive data systems to be as near real-time as possible.
- Support design and deployment of a distributed data store that will be the central source of truth across the organization.
- Build data connections to the company's internal IT systems.
- Develop, customize, and configure self-service tools that help data consumers extract and analyze data from our massive internal data store.
- Evaluate new technologies and build prototypes for continuous improvements in data engineering.
What We're Looking For
- 5+ years of work experience in a relevant field (Data Engineer, DWH Engineer, Software Engineer, etc).
- Experience with data-lake and data-warehousing technologies and relevant data modeling best practices (Presto, Athena, Glue, etc).
- Proficiency in at least one of the main programming languages used: Python and Scala.
- Experience building data pipelines/ETL in Airflow, and familiarity with software design principles.
- Excellent SQL and data manipulation skills using common frameworks like Spark/PySpark, or similar.
- Expertise in Apache Spark, or similar Big Data technologies, with a proven record of processing high volumes and velocity of datasets.
- Experience with business requirements gathering for data sourcing.
Nice to Have
- Additional programming languages expertise.
- Experience with Kafka and other streaming technologies like Apache Flink.
Technical Stack
- Languages: Python, Scala, SQL
- Big Data: Apache Spark, PySpark
- Orchestration: Airflow
- Query Engines & Data Services: Presto, Athena, Glue
- Streaming: Kafka, Apache Flink
Team & Environment
You will be part of the Data Engineering organization, working specifically on the Data Platform team.
Work Mode
This is a global role, open to candidates in 70+ countries.
As an equal opportunity employer, we don’t tolerate discrimination or harassment of any kind. Whether that’s based on race, ethnicity, age, gender identity, citizenship, religion, sexual orientation, disability, pregnancy, veteran status or any other protected characteristic as outlined by federal, state or local laws.





