J.P. Morgan
AWS Software Engineer III-Databricks/Python/Pyspark
J.P. Morgan, Wilmington, Delaware, us, 19894
We have an exciting and rewarding opportunity for you to take your software engineering career to the next level.
As an AWS Software Engineer III-Databricks/Python/PySpark at JPMorgan Chase within the Corporate Sector-Global Finance Team, you will be a key member of an agile team, tasked with designing and delivering cutting‑edge products that are secure, stable, and scalable. Your role involves implementing essential technology solutions across diverse technical domains to support the firm's business goals effectively.
Job responsibilities
Developing and optimizing data pipelines and workflows to support data integration, transformation, and analysis
Implementing best practices for data management, ensuring data quality, security, and compliance
Writing secure, high‑quality production code following AWS best practices, and deploying efficiently using CI/CD pipelines
Creating architecture and design documents for complex applications, ensuring software code meets design constraints
Identifying hidden issues and patterns in data to enhance coding practices and system architecture
Contributing to software engineering communities, promoting diversity, opportunity, inclusion, and respect within the team
Required qualifications, capabilities, and skills
Formal training or certification on software engineering concepts and 3+ years applied experience
Experience with Spark and SQL
Expertise in Lakehouse/Delta Lake architecture, system design, application development, testing, and ensuring operational stability
Strong programming skills in Python/PySpark
Proficient in orchestration using Airflow
In‑depth knowledge of Big Data and data warehousing concepts
Proficient in SQL/SparkSQL
Experience with CI/CD processes
Thorough understanding of the Software Development Life Cycle (SDLC)
Solid understanding of agile methodologies, including DevOps practices, application resiliency, and security measures
Proven expertise in software applications and technical processes within a specialized technical domain
Preferred qualifications, capabilities, and skills
Experience in full‑stack development with strong proficiency in Python
Experience with Databricks, and the AWS cloud ecosystem
Familiarity with Snowflake, Terraform and LLM
Exposure to cloud technologies such as AWS Glue, S3, SQS/SNS, Lambda etc.
Familiarity with Data Observability, Data Quality, Query Optimization & Cost Optimization
AWS certifications such as SAA, Associate Developer, Data Analytics Specialty, or Databricks certification
#J-18808-Ljbffr
As an AWS Software Engineer III-Databricks/Python/PySpark at JPMorgan Chase within the Corporate Sector-Global Finance Team, you will be a key member of an agile team, tasked with designing and delivering cutting‑edge products that are secure, stable, and scalable. Your role involves implementing essential technology solutions across diverse technical domains to support the firm's business goals effectively.
Job responsibilities
Developing and optimizing data pipelines and workflows to support data integration, transformation, and analysis
Implementing best practices for data management, ensuring data quality, security, and compliance
Writing secure, high‑quality production code following AWS best practices, and deploying efficiently using CI/CD pipelines
Creating architecture and design documents for complex applications, ensuring software code meets design constraints
Identifying hidden issues and patterns in data to enhance coding practices and system architecture
Contributing to software engineering communities, promoting diversity, opportunity, inclusion, and respect within the team
Required qualifications, capabilities, and skills
Formal training or certification on software engineering concepts and 3+ years applied experience
Experience with Spark and SQL
Expertise in Lakehouse/Delta Lake architecture, system design, application development, testing, and ensuring operational stability
Strong programming skills in Python/PySpark
Proficient in orchestration using Airflow
In‑depth knowledge of Big Data and data warehousing concepts
Proficient in SQL/SparkSQL
Experience with CI/CD processes
Thorough understanding of the Software Development Life Cycle (SDLC)
Solid understanding of agile methodologies, including DevOps practices, application resiliency, and security measures
Proven expertise in software applications and technical processes within a specialized technical domain
Preferred qualifications, capabilities, and skills
Experience in full‑stack development with strong proficiency in Python
Experience with Databricks, and the AWS cloud ecosystem
Familiarity with Snowflake, Terraform and LLM
Exposure to cloud technologies such as AWS Glue, S3, SQS/SNS, Lambda etc.
Familiarity with Data Observability, Data Quality, Query Optimization & Cost Optimization
AWS certifications such as SAA, Associate Developer, Data Analytics Specialty, or Databricks certification
#J-18808-Ljbffr