About the Role
The role involves building and optimizing data pipelines, working closely with analytics and engineering teams to support scalable data solutions.
Responsibilities
- Develop and manage scalable data pipelines
- Ensure data accuracy and reliability across systems
- Collaborate with data analysts and scientists
- Optimize data storage and retrieval processes
- Support data governance and quality standards
- Monitor performance of data infrastructure
- Troubleshoot and resolve data issues
- Implement data models for reporting and analytics
- Work with cloud-based data platforms
- Document data workflows and system designs
- Integrate data from multiple sources
- Maintain data security and compliance
- Automate routine data operations
- Contribute to data architecture planning
- Support business intelligence initiatives
- Participate in code reviews
- Improve data processing efficiency
- Assist in database administration
- Use version control for data pipeline code
- Engage in agile project cycles
- Respond to data access requests
- Evaluate new data tools and technologies
- Ensure system reliability and uptime
- Collaborate on data-driven product features
- Promote best practices in data engineering
Nice to Have
- Master's degree in a technical field
- Experience in education technology sector
- Knowledge of machine learning pipelines
- Familiarity with containerization tools
- Experience with data governance frameworks
- Exposure to real-time data processing
- Background in data security
- Certifications in cloud platforms
- Public speaking at tech events
- Open-source contributions
Compensation
Competitive salary with benefits
Work Arrangement
Hybrid work model
Team
Part of a growing data team within a global organization focused on education technology
Why Join Us
- Opportunity to shape the future of data infrastructure
- Work in a mission-driven environment focused on global education access
- Collaborative culture with emphasis on innovation
- Support for professional development and conferences
- Inclusive workplace with diverse teams
Our Tech Stack
- Cloud infrastructure on Google Cloud Platform
- Data pipelines using Apache Airflow
- Primary language is Python
- Data warehouse in BigQuery
- Version control via Git
- Containerization with Docker
- CI/CD through GitLab CI
- Monitoring with Prometheus and Grafana
- SQL for analytics
- Kafka for event streaming
Application Process
- Submit resume and cover letter
- Initial screening call
- Technical assessment
- Interview with team members
- Final discussion with hiring manager
- Reference check
- Offer and onboarding
Available for qualified candidates