Responsibilities
- Lead data architecture design, API assessment, and ETL requirements gathering during the Discovery & Design phase.
- Develop and configure CMIC ERP API integration to establish reliable data exchange between the ERP system and the AWS platform.
- Design and implement data pipelines using AWS Glue for ETL processing of subcontractor documents and ERP data.
- Integrate Amazon Textract to extract structured data from insurance certificates, bonding letters, and financial documents.
- Build and maintain data models to support AI-powered validation, risk profiling, and executive reporting.
- Configure Amazon S3 data lake architecture to store and manage raw, processed, and curated data assets.
- Implement AWS Lambda and AWS Step Functions to orchestrate data workflows and automated processing pipelines.
- Develop and expose data through Amazon API Gateway to support application and dashboard consumption.
- Ensure data quality, validation, and integrity across all integration points and pipeline outputs.
- Conduct data integration testing and support user acceptance testing (UAT) for data-dependent features.
- Collaborate with Full Stack, AI/ML, and DevOps team members to ensure seamless end-to-end data flows.
- Contribute to knowledge transfer documentation, data pipeline runbooks, and operations guides.
Requirements
- 5+ years of data engineering experience, with at least 2+ years working in AWS cloud environments.
- Strong proficiency in SQL and experience with AWS database services including RDS, DynamoDB, and Aurora.
- Hands-on experience with AWS Glue for ETL development and data pipeline orchestration.
- Experience integrating with ERP systems or enterprise APIs, with ability to assess and document API capabilities.
- Working knowledge of Amazon Textract for automated document data extraction.
- Proficiency with Amazon S3 for data lake design, storage management, and lifecycle policies.
- Experience with AWS Lambda and AWS Step Functions for serverless data workflow orchestration.
- Solid understanding of data modeling, data quality validation, and integration testing practices.
- Familiarity with Amazon API Gateway for exposing data services to downstream consumers.
- Strong analytical, problem-solving, and communication skills with the ability to work in Agile/Scrum teams.
Nice to Have
- Experience with Amazon Bedrock or AI/ML-integrated data pipelines.
- Familiarity with CMIC ERP or similar construction industry enterprise resource planning systems.
- Experience with GraphQL APIs or AWS AppSync.
- Knowledge of TypeScript or Python for pipeline scripting and Lambda function development.
- AWS Certification (Data Analytics Specialty, Database Specialty, or Solutions Architect).
- Background in construction, insurance, or financial services industries.
Benefits
- Remote work
Additional Information
- 1099
- No Agencies Please!
- Capnexus is an equal opportunity employer.
- We embrace and celebrate diversity and are committed to creating an inclusive and safe environment for all employees.
- We encourage you to apply even if your experience doesn’t perfectly align with what we have listed.