Responsibilities
- Build and maintain scalable data pipelines that ingest and transform financial, education, and employment data
- Ensure data is reliable, timely, and accessible across internal teams and external partners
- Proactively identify and resolve data quality issues, including upstream dependencies
- Partner with stakeholders in finance, operations and business development departments to ensure that their data sets are reliably ingested into our data warehouse and their questions or reports can be answered programmatically
- Ensure via automated testing and edge case handling that we can detect any errors with upstream data or data processing and that all reports contain the data as expected
- Develop processes to anonymize and protect sensitive data across environments
- Support the configuration and optimization of data warehouse, storage, and compute resources
- Partner in managing infrastructure as code (e.g., Terraform) to ensure reproducibility and scalability
- Monitor performance and cost efficiency of data workloads, identifying opportunities for optimization
- Contribute to improving the reliability, scalability, and observability of our data platform
Requirements
- 3+ years of experience as a data engineer, or equivalent experience building data pipelines
- 2+ years of experience with elements of the following technologies:
- Data Analytics: Airflow, DBT
- Business Intelligence Systems: Tableau, Looker, Sisense, Apache Superset, etc.
- Languages: Python, SQL
- Databases: SQL; NoSQL a bonus
- Strong communication and collaboration skills, with the ability to effectively communicate the complexities of technical programs to both technical and nontechnical stakeholders
- Desire to mentor and collaborate with other members of the team
- Willingness to roll your sleeves up to rapidly acquire competencies in a wide range of technical disciplines
- Bachelor’s degree in Computer Science, Software Engineering, Information Systems, or equivalent experience
Nice to Have
- Data Warehousing: Snowflake, BigQuery, etc.
- Cloud Infrastructure: AWS or Google Cloud Experience
- DevOps: Experience with modern cloud and container tooling such as Docker, Kubernetes, Terraform, etc.
Benefits
- Equity component as part of our compensation package, providing an opportunity for eligible employees to share in the success and growth of our company.
Work Arrangement
Hybrid
Additional Information
- To ensure smooth collaboration with the Boston-based team, we are limiting to Eastern/Central timezones with the expectation that it will be Boston hours. Additionally, they must be within 1 hour of a medium hub/ large airport with multiple daily flights to Boston. We are targeting the broader New England area to start as we expect that the candidate to have frequent travel to Boston for onboarding and ongoing collaboration.