Capstone Integrated Solutions Logo

Capstone Integrated Solutions

Senior Data Engineer (AWS)

Posted 9 Days Ago
Remote
Hiring Remotely in USA
Senior level
Remote
Hiring Remotely in USA
Senior level
Lead data architecture, develop ETL pipelines using AWS Glue, integrate with CMIC ERP, and ensure data quality for AI-driven projects.
The summary above was generated by AI

Capnexus is a comprehensive services provider. Our team consists of outstanding professionals, highly experienced in designing, building, and supporting retail software. We see ourselves as a build-as-a-service provider who follows a repeatable business pattern that can be applied to a variety of platforms and verticals. Having a culture built on outcomes and delivery at the core of the business, Capnexus is providing its customers with a complete suite of services for software development, system analysis, integration, implementation, and support, as well as the option to engage a single team to perform all the services they require. 

Who You Are and What You'll Do: 

Capnexus is looking for a highly skilled Senior AWS Data Engineer to lead data architecture, pipeline development, and ERP integration for a 12-week AI-powered modernization engagement in the construction industry. This role is focused on designing and implementing the data engineering backbone of an intelligent subcontractor pre-qualification platform, including CMIC ERP API integration, Amazon Textract data extraction pipelines, ETL development using AWS Glue, and data quality validation. This is an exciting opportunity to apply advanced cloud data engineering skills on a platform that leverages generative AI to automate and modernize enterprise workflows. 

Responsibilities: 

  • Lead data architecture design, API assessment, and ETL requirements gathering during the Discovery & Design phase. 
  • Develop and configure CMIC ERP API integration to establish reliable data exchange between the ERP system and the AWS platform. 
  • Design and implement data pipelines using AWS Glue for ETL processing of subcontractor documents and ERP data. 
  • Integrate Amazon Textract to extract structured data from insurance certificates, bonding letters, and financial documents. 
  • Build and maintain data models to support AI-powered validation, risk profiling, and executive reporting. 
  • Configure Amazon S3 data lake architecture to store and manage raw, processed, and curated data assets. 
  • Implement AWS Lambda and AWS Step Functions to orchestrate data workflows and automated processing pipelines. 
  • Develop and expose data through Amazon API Gateway to support application and dashboard consumption. 
  • Ensure data quality, validation, and integrity across all integration points and pipeline outputs. 
  • Conduct data integration testing and support user acceptance testing (UAT) for data-dependent features. 
  • Collaborate with Full Stack, AI/ML, and DevOps team members to ensure seamless end-to-end data flows. 
  • Contribute to knowledge transfer documentation, data pipeline runbooks, and operations guides. 

 

Qualifications: 

  • 5+ years of data engineering experience, with at least 2+ years working in AWS cloud environments. 
  • Strong proficiency in SQL and experience with AWS database services including RDS, DynamoDB, and Aurora. 
  • Hands-on experience with AWS Glue for ETL development and data pipeline orchestration. 
  • Experience integrating with ERP systems or enterprise APIs, with ability to assess and document API capabilities. 
  • Working knowledge of Amazon Textract for automated document data extraction. 
  • Proficiency with Amazon S3 for data lake design, storage management, and lifecycle policies. 
  • Experience with AWS Lambda and AWS Step Functions for serverless data workflow orchestration. 
  • Solid understanding of data modeling, data quality validation, and integration testing practices. 
  • Familiarity with Amazon API Gateway for exposing data services to downstream consumers. 
  • Strong analytical, problem-solving, and communication skills with the ability to work in Agile/Scrum teams. 

 

Nice to Have: 

  • Experience with Amazon Bedrock or AI/ML-integrated data pipelines. 
  • Familiarity with CMIC ERP or similar construction industry enterprise resource planning systems. 
  • Experience with GraphQL APIs or AWS AppSync. 
  • Knowledge of TypeScript or Python for pipeline scripting and Lambda function development. 
  • AWS Certification (Data Analytics Specialty, Database Specialty, or Solutions Architect). 
  • Background in construction, insurance, or financial services industries. 

 

"Our Culture": 

At Capstone, the central principles that we all adhere to, and the glue that holds us together, are our keystones. Our four keystones are: 

"A Customer Obsessed, Delivery Focused, Culture" 

  • We’re driven to exceed our customers’ expectations by listening, leading, solving problems, and delivering what we promise 
  • We aim to be the most dependable and trusted partner serving our customers. TRUST = CONSISTENCY x TIME 

"A Culture of Learning and Sharing" 

  • We value “Lifetime Learners”; those who are hungry, competitive, curious, and self-motivated in their pursuit of knowledge. 
  • Personal and professional growth depends on teamwork and continuous learning. By sharing knowledge, skills, ideas, and effort, we benefit our customers, ourselves, and our communities. 
  • We recognize that the thoughts, feelings, and backgrounds of others are as important as our own. Everyone has something to learn and everyone has something they can teach. 
  • Knowledge and ability are valued. Sharing knowledge and helping others learn new capabilities is valued exponentially. 

"A Culture of Growth and Scalability" 

  • Growth comes from not establishing barriers in your role. “Cross functional skill sets are valued and help us deliver to our customers in a truly agile fashion. It comes with understanding that when asked to do something new, you will need support, have questions, and make some mistakes along the way. 
  • The most elegant solution is a simple solution. Simple doesn’t mean easy. It’s often more difficult to break a complex problem down into simple, scalable terms. We don’t appreciate, or value, over architected solutions or superfluous coding. 
  • Time is one of our most precious commodities. Scalability implies being respectful of this and passionate about making the most efficient use of each and every one of our team members time. 

"All Work is Strategic" 

  • No matter how small a project or assignment appears, every single engagement is an opportunity for us to prove ourselves, build trust, and develop relationships that last and grow 
  • Every task, interaction, and commitment matters 
  • Big or small, we execute our plans and strategies with focus, commitment, and passion 

 

We offer: 

Job Type: Full-time, 1099

Duration: Short-term

Benefits: 

  • Remote work 

 

Capnexus is an equal opportunity employer. We embrace and celebrate diversity and are committed to creating an inclusive and safe environment for all employees. Experience comes in many forms, and we’re dedicated to adding new perspectives to the team. We encourage you to apply even if your experience doesn’t perfectly align with what we have listed. We look forward to hearing from you. 

No Agencies Please! 

Similar Jobs

4 Days Ago
In-Office or Remote
Senior level
Senior level
Information Technology • Software
The Senior Data Platform Engineer leads the design and management of data pipelines using Python and AWS, focusing on data engineering for finance-related analytics and systems.
Top Skills: AirflowAWSCdkCloudFormationDockerEcsEventbridgeFastapiFlaskGlueKinesisKubernetesLambdaNumpyPandasPrefectPysparkPytestPythonS3SqlalchemyStep FunctionsTerraform
19 Days Ago
Remote
USA
Senior level
Senior level
Artificial Intelligence • Information Technology • Software • Database
Build production-grade AI-powered data tooling: extract data from Snowflake, generate and store embeddings, enable semantic search, design enrichment pipelines using LLM APIs, optimize AWS infrastructure, and create reusable services and SDKs for scalable, observable data and AI workflows.
Top Skills: AWSEcsEksEmbeddingsIamLambdaLlm ApisOpenaiPineconePythonS3Semantic SearchSnowflakeVector Databases
24 Days Ago
In-Office or Remote
Seattle, WA, USA
Mid level
Mid level
Database
As an AWS Data Engineer, you will develop data pipelines, create data models, troubleshoot issues, and collaborate with teams while utilizing tools like PySpark and SQL.
Top Skills: AirflowAmazon AthenaAmazon EmrAmazon GlueAmazon RedshiftPrestoPysparkSnowflakeSQL

What you need to know about the Seattle Tech Scene

Home to tech titans like Microsoft and Amazon, Seattle punches far above its weight in innovation. But its surrounding mountains, sprinkled with world-famous hiking trails and climbing routes, make the city a destination for outdoorsy types as well. Established as a logging town before shifting to shipbuilding and logistics, the Emerald City is now known for its contributions to aerospace, software, biotech and cloud computing. And its status as a thriving tech ecosystem is attracting out-of-town companies looking to establish new tech and engineering hubs.

Key Facts About Seattle Tech

  • Number of Tech Workers: 287,000; 13% of overall workforce (2024 CompTIA survey)
  • Major Tech Employers: Amazon, Microsoft, Meta, Google
  • Key Industries: Artificial intelligence, cloud computing, software, biotechnology, game development
  • Funding Landscape: $3.1 billion in venture capital funding in 2024 (Pitchbook)
  • Notable Investors: Madrona, Fuse, Tola, Maveron
  • Research Centers and Universities: University of Washington, Seattle University, Seattle Pacific University, Allen Institute for Brain Science, Bill & Melinda Gates Foundation, Seattle Children’s Research Institute

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account