HomePython Backend DeveloperPython Backend Developer with AWS
IT America Inc

Python Backend Developer with AWS

IT America Inc·McLean, Virginia, US

Posted 1w ago

Full-Time
Apply Now

About the Role

Position: Python Backend Developer with AWS Location: McLean, VA (Preferred) | Richmond, VA | Plano, TX Duration: Long term contract Note: Ex–Capital One candidates preferred, especially those able to work at client locations. Role Overview: Capital One is seeking a skilled Backend Developer with strong Data Engineering expertise to design, build, and maintain scalable data pipelines in a cloud-native environment. This role focuses on backend development using Python, PySpark, and AWS, supporting high-volume data processing and analytics platforms. The ideal candidate has hands-on experience building production-grade data pipelines, working with distributed data frameworks, and deploying solutions on AWS. Key Responsibilities • Design, develop, and maintain backend data pipelines using Python and PySpark • Build and optimize ETL/ELT workflows for large-scale data processing • Develop cloud-hosted solutions using AWS services • Ensure performance, reliability, and scalability of data engineering pipelines • Collaborate with product owners, data scientists, and platform teams to support analytics and business use cases • Implement data quality checks, logging, monitoring, and error handling • Participate in code reviews and follow Capital One engineering best practices • Support deployment and operations in a CI/CD-driven cloud environment Required Skills & Qualifications • Strong hands-on experience with Python backend development • Solid experience using PySpark for large-scale data processing • Experience hosting and deploying applications on AWS (e.g., S3, EC2, EMR, Glue, Lambda, CloudWatch – exact services may vary) • Proven experience building data engineering pipelines • Understanding of distributed systems and data processing concepts • Experience working in Agile/Scrum development environments • Strong problem-solving and debugging skills Preferred Qualifications • Ex–Capital One experience (highly preferred) • Ability to work onsite in McLean, VA Experience with: • CI/CD pipelines • Data lake architectures • Cloud-native backend systems • Financial services or enterprise-scale data platform experience Nice-to-Have Skills • SQL and data modeling • Workflow orchestration tools (Airflow or similar) • Exposure to streaming or near–real-time data processing • Knowledge of security and compliance in cloud environments

What you'll do

  • Capital One is seeking a skilled Backend Developer with strong Data Engineering expertise to design, build, and maintain scalable data pipelines in a cloud-native environment
  • This role focuses on backend development using Python, PySpark, and AWS, supporting high-volume data processing and analytics platforms
  • Design, develop, and maintain backend data pipelines using Python and PySpark
  • Build and optimize ETL/ELT workflows for large-scale data processing
  • Develop cloud-hosted solutions using AWS services
  • Ensure performance, reliability, and scalability of data engineering pipelines
  • Collaborate with product owners, data scientists, and platform teams to support analytics and business use cases
  • Implement data quality checks, logging, monitoring, and error handling
  • Participate in code reviews and follow Capital One engineering best practices
  • Support deployment and operations in a CI/CD-driven cloud environment

Requirements

  • The ideal candidate has hands-on experience building production-grade data pipelines, working with distributed data frameworks, and deploying solutions on AWS
  • Strong hands-on experience with Python backend development
  • Solid experience using PySpark for large-scale data processing
  • Experience hosting and deploying applications on AWS (e.g., S3, EC2, EMR, Glue, Lambda, CloudWatch – exact services may vary)
  • Proven experience building data engineering pipelines
  • Understanding of distributed systems and data processing concepts
  • Experience working in Agile/Scrum development environments
  • Strong problem-solving and debugging skills
  • CI/CD pipelines
  • Data lake architectures
  • Cloud-native backend systems
  • Financial services or enterprise-scale data platform experience
  • SQL and data modeling
  • Workflow orchestration tools (Airflow or similar)
  • Exposure to streaming or near–real-time data processing
  • Knowledge of security and compliance in cloud environments
Back to all jobs