Hybrid Full-time

Granica is hiring a Lakehouse Engineer

About the Role

Granica is hiring a Lakehouse Engineer to design and implement robust, scalable data pipelines from scratch in mission-critical environments. You’ll configure modern data lakehouse technologies and act as a trusted technical advisor guiding customers through architecture, deployment, and troubleshooting.

What You'll Do

  • Partner closely with customers to understand their technical environment, data challenges, and integration needs.
  • Design and implement robust, scalable data pipelines from scratch using PySpark and Python.
  • Configure and integrate modern data lakehouse and warehouse technologies including Apache Iceberg, Apache Hive, Delta Lake, Snowflake, and Databricks.
  • Act as a trusted technical advisor, guiding customers through solution architecture, deployment, and troubleshooting.
  • Contribute to internal tooling and automation to improve deployment velocity and system reliability.
  • Collaborate with Granica’s engineering and product teams to influence roadmap decisions based on real-world customer use cases.
  • Be an ambassador of the Granica product, both internally and externally.

What We're Looking For

  • 5–10 years of hands-on experience in software engineering, data engineering, or infrastructure roles.
  • Strong proficiency in Python and PySpark, with the ability to write clean, efficient, and scalable code.
  • Proven experience building data pipelines from scratch, including ingestion, transformation, and optimization.
  • Deep understanding and hands-on experience with: Apache Iceberg, Apache Hive, Apache Delta Lake, Snowflake, Databricks.
  • Experience working with large-scale data systems and distributed computing architectures.
  • Ability to thrive in fast-paced, ambiguous environments typical of early-stage startups.
  • Excellent problem-solving, communication, and customer-facing skills.

Nice to Have

  • Experience with Kubernetes, Terraform, or cloud-native infrastructure (AWS/GCP/Azure).
  • Familiarity with security and privacy best practices in data processing pipelines.
  • Prior experience in customer-facing technical roles (solutions engineer, customer engineer, etc.) is a strong plus.

Technical Stack

  • Languages & Frameworks: Python, PySpark
  • Lakehouse Technologies: Apache Iceberg, Apache Hive, Delta Lake, Snowflake, Databricks
  • Infrastructure: Kubernetes, Terraform, AWS, GCP, Azure

Benefits & Compensation

  • Highly competitive compensation + meaningful equity
  • Unlimited PTO
  • Flexible hybrid setup
  • Full health coverage including premium health, dental, and vision
  • Quarterly recharge days
  • Annual team off-sites
  • Budget for learning, development, and conferences
  • Immigration sponsorship and counseling

Work Mode

This role operates on a hybrid work model.

Granica celebrates diversity and is committed to creating an inclusive environment for all employees. We do not discriminate on the basis of race, religion, color, gender expression or identity, sexual orientation, national origin, citizenship, age, marital status, veteran status, disability status, or any other characteristic protected by law.

Required Skills
PythonPySparkApache IcebergApache HiveDelta LakeSnowflakeDatabricksKubernetesTerraformAWSData EngineeringLakehouse ArchitectureCloud Infrastructure
Invoicing holding you back?

Focus on work, not paperwork

Stop worrying about invoicing, taxes, and compliance. Glopay handles the business setup, you handle the client work. Get paid faster and look professional.

Auto-generated compliant invoices
Built-in expense management
Income reports for tax season
95% of earnings stay with you
Try Glopay free
No credit card needed
About company
Granica

Granica is an AI research and systems company building infrastructure for intelligence that is structured, efficient, and deeply integrated with data, operating at exabyte scale and processing petabytes of data daily for prominent enterprises.

Visit website
Job Details
Category data
Posted 7 months ago