Logo
J.P. Morgan

AWS Software Engineer III-Databricks/Python/Pyspark

J.P. Morgan, Wilmington, Delaware, us, 19894

Save Job

We have an exciting and rewarding opportunity for you to take your software engineering career to the next level.

As an AWS Software Engineer III-Databricks/Python/PySpark at JPMorgan Chase within the Corporate Sector-Global Finance Team, you will be a key member of an agile team, tasked with designing and delivering cutting‑edge products that are secure, stable, and scalable. Your role involves implementing essential technology solutions across diverse technical domains to support the firm's business goals effectively.

Job responsibilities

Developing and optimizing data pipelines and workflows to support data integration, transformation, and analysis

Implementing best practices for data management, ensuring data quality, security, and compliance

Writing secure, high‑quality production code following AWS best practices, and deploying efficiently using CI/CD pipelines

Creating architecture and design documents for complex applications, ensuring software code meets design constraints

Identifying hidden issues and patterns in data to enhance coding practices and system architecture

Contributing to software engineering communities, promoting diversity, opportunity, inclusion, and respect within the team

Required qualifications, capabilities, and skills

Formal training or certification on software engineering concepts and 3+ years applied experience

Experience with Spark and SQL

Expertise in Lakehouse/Delta Lake architecture, system design, application development, testing, and ensuring operational stability

Strong programming skills in Python/PySpark

Proficient in orchestration using Airflow

In‑depth knowledge of Big Data and data warehousing concepts

Proficient in SQL/SparkSQL

Experience with CI/CD processes

Thorough understanding of the Software Development Life Cycle (SDLC)

Solid understanding of agile methodologies, including DevOps practices, application resiliency, and security measures

Proven expertise in software applications and technical processes within a specialized technical domain

Preferred qualifications, capabilities, and skills

Experience in full‑stack development with strong proficiency in Python

Experience with Databricks, and the AWS cloud ecosystem

Familiarity with Snowflake, Terraform and LLM

Exposure to cloud technologies such as AWS Glue, S3, SQS/SNS, Lambda etc.

Familiarity with Data Observability, Data Quality, Query Optimization & Cost Optimization

AWS certifications such as SAA, Associate Developer, Data Analytics Specialty, or Databricks certification

#J-18808-Ljbffr