Stripe
Overview
Staff Software Engineer, Data Movement role at Stripe. The Data Movement team builds and operates a large-scale read-only ecosystem extending Stripes primary online database, enabling read and query capabilities at massive scale. The team processes real-time ingestion of terabytes of data and billions of events into the Data Warehouse for analysis and business-specific processing, operating some of Stripes highest-volume data pipelines. Key technologies include Flink, Spark (structured streaming), Iceberg, Trino, Airflow, Kafka, CDC, and Kubernetes, across multiple regions. The role is high-visibility and senior, providing technical direction and serving as a culture north star for the organization. What youll do
Design, build, and maintain innovative solutions with an emphasis on reliability, scalability, and efficiency. Define architectural strategies, policies, and abstractions for long-term, sustainable internal Stripe solutions that improve the experience of millions of Stripe users. Ensure operational excellence and reliability of the Data Movement platform across streaming and batch workloads, delivering a world-class data and query ecosystem. Collaborate with high-visibility teams and stakeholders to support key initiatives, navigating organizational complexity with leadership support. Contribute to usability and reliability improvements of OSS like Apache Airflow, Iceberg, Spark, and contribute back to the OSS community. Who you are
Minimum Requirements
Strong experience with large-scale, high-availability data pipelines; design, execution, and safe change management. Proficiency in Flink, Spark (structured streaming), Airflow, Python, Java, SQL, Trino, and API design is a plus. Experience developing, maintaining, and debugging distributed systems built with open-source tools. Experience building infrastructure-as-a-product with a strong focus on user needs. Strong collaboration and communication skills; able to interact with both technical and non-technical participants. Curiosity to continuously learn about new technologies and business processes. Energy to deliver effective, user-first solutions through creative problem-solving and collaboration. Preferred Qualifications
Production-level code experience in Scala, Spark, Flink, Iceberg, Airflow, Python, Java, and SQL is a plus. Experience packaging and deploying code into cloud environments (AWS, GCP, Azure) with tools like Bazel and Docker. Experience designing APIs or building developer platforms. Experience optimizing end-to-end performance of distributed systems. Experience scaling distributed systems in a fast-moving environment. Experience with data pipelines. Genuine enjoyment of innovation and a deep interest in understanding how things work. Work model
Hybrid work at Stripe: this role is available in an office or remote location (35+ miles / 56+ km from a Stripe office). In-office and remote expectations are described by Stripes policies. #J-18808-Ljbffr
Staff Software Engineer, Data Movement role at Stripe. The Data Movement team builds and operates a large-scale read-only ecosystem extending Stripes primary online database, enabling read and query capabilities at massive scale. The team processes real-time ingestion of terabytes of data and billions of events into the Data Warehouse for analysis and business-specific processing, operating some of Stripes highest-volume data pipelines. Key technologies include Flink, Spark (structured streaming), Iceberg, Trino, Airflow, Kafka, CDC, and Kubernetes, across multiple regions. The role is high-visibility and senior, providing technical direction and serving as a culture north star for the organization. What youll do
Design, build, and maintain innovative solutions with an emphasis on reliability, scalability, and efficiency. Define architectural strategies, policies, and abstractions for long-term, sustainable internal Stripe solutions that improve the experience of millions of Stripe users. Ensure operational excellence and reliability of the Data Movement platform across streaming and batch workloads, delivering a world-class data and query ecosystem. Collaborate with high-visibility teams and stakeholders to support key initiatives, navigating organizational complexity with leadership support. Contribute to usability and reliability improvements of OSS like Apache Airflow, Iceberg, Spark, and contribute back to the OSS community. Who you are
Minimum Requirements
Strong experience with large-scale, high-availability data pipelines; design, execution, and safe change management. Proficiency in Flink, Spark (structured streaming), Airflow, Python, Java, SQL, Trino, and API design is a plus. Experience developing, maintaining, and debugging distributed systems built with open-source tools. Experience building infrastructure-as-a-product with a strong focus on user needs. Strong collaboration and communication skills; able to interact with both technical and non-technical participants. Curiosity to continuously learn about new technologies and business processes. Energy to deliver effective, user-first solutions through creative problem-solving and collaboration. Preferred Qualifications
Production-level code experience in Scala, Spark, Flink, Iceberg, Airflow, Python, Java, and SQL is a plus. Experience packaging and deploying code into cloud environments (AWS, GCP, Azure) with tools like Bazel and Docker. Experience designing APIs or building developer platforms. Experience optimizing end-to-end performance of distributed systems. Experience scaling distributed systems in a fast-moving environment. Experience with data pipelines. Genuine enjoyment of innovation and a deep interest in understanding how things work. Work model
Hybrid work at Stripe: this role is available in an office or remote location (35+ miles / 56+ km from a Stripe office). In-office and remote expectations are described by Stripes policies. #J-18808-Ljbffr