phData, a leader in the modern data stack, is looking for a Data Engineer. You will be responsible for developing end-to-end technical solutions into production, ensuring performance, security, scalability, and robust data integration.
What You'll Do
- Develop end-to-end technical solutions into production
- Help ensure performance, security, scalability, and robust data integration
- Create and deliver detailed presentations
- Create detailed solution documentation (e.g., including POCs and roadmaps, sequence diagrams, class hierarchies, logical system views, etc.)
What We're Looking For
- At least 4+ years experience as a Software Engineer, Data Engineer or Data Analyst
- Ability to develop end-to-end technical solutions into production
- Programming expertise in Java, Python and/or Scala
- Experience with core cloud data platforms including Snowflake, AWS, Azure, Databricks and GCP
- SQL and the ability to write, debug, and optimize SQL queries
- Client-facing English written and verbal communication skills and experience
- 4-year Bachelor's degree in Computer Science or a related field
- Strong English communication skills (written and verbal)
Nice to Have
- Production experience in core data platforms: Snowflake, AWS, Azure, GCP, Hadoop, Databricks
- Experience with Cloud and Distributed Data Storage: S3, ADLS, HDFS, GCS, Kudu, ElasticSearch/Solr, Cassandra or other NoSQL storage systems
- Experience with Data integration technologies: Spark, Kafka, event/streaming, Streamsets, Matillion, Fivetran, NiFi, AWS Data Migration Services, Azure DataFactory, Informatica Intelligent Cloud Services (IICS), Google DataProc or other data integration technologies
- Experience with Multiple data sources (e.g., queues, relational databases, files, search, API)
- Complete software development lifecycle experience including design, documentation, implementation, testing, and deployment
- Experience with Automated data transformation and data curation: dbt, Spark, Spark streaming, automated pipelines
- Experience with Workflow Management and Orchestration: Airflow, AWS Managed Airflow, Luigi, NiFi
Technical Stack
- Languages: Java, Python, Scala
- Core Platforms: Snowflake, AWS, Azure, Databricks, GCP, Hadoop
- Storage: S3, ADLS, HDFS, GCS, Kudu, ElasticSearch/Solr, Cassandra
- Integration & Streaming: Spark, Kafka, Streamsets, Matillion, Fivetran, NiFi, AWS Data Migration Services, Azure DataFactory, Informatica Intelligent Cloud Services (IICS), Google DataProc
- Transformation & Orchestration: dbt, Airflow, AWS Managed Airflow, Luigi
Benefits & Compensation
- Remote-First Work Environment
- Casual, award-winning small-business work environment
- Collaborative culture that prizes autonomy, creativity, and transparency
- Competitive comp, excellent benefits, generous PTO plan plus 10 Holidays
- Accelerated learning and professional development through advanced training and certifications
Work Mode
This is a remote-first global position open to candidates in the United States, Latin America, and India.
phData celebrates diversity and is committed to creating an inclusive environment for all employees. We are proud to be an equal opportunity employer. We prohibit discrimination and harassment of any kind based on race, color, religion, national origin, sex (including pregnancy), sexual orientation, gender identity, gender expression, age, veteran status, genetic information, disability, or other applicable legally protected characteristics.



