Logo
Roku, Inc.

Senior Software Engineer, DevOps - Data Platform

Roku, Inc., Santa Monica, California, United States, 90403

Save Job

Teamwork makes the stream work. Roku is changing how the world watches TV Roku is the #1 TV streaming platform in the U.S., Canada, and Mexico, and we've set our sights on powering every television in the world. Roku pioneered streaming to the TV. Our mission is to be the TV streaming platform that connects the entire TV ecosystem. We connect consumers to the content they love, enable content publishers to build and monetize large audiences, and provide advertisers unique capabilities to engage consumers.

From your first day at Roku, you'll make a valuable - and valued - contribution. We're a fast-growing public company where no one is a bystander. We offer you the opportunity to delight millions of TV streamers around the world while gaining meaningful experience across a variety of disciplines.

About the Team Roku runs one of the largest data lakes in the world. We store over 70 petabytes of data, run more than 10 million queries per month, and scan over 100 petabytes of data per month. The Big Data team is the one responsible for building, running, and supporting the platform that makes this possible. We provide all the necessary tooling to acquire, generate, process, monitor, validate, and access data in the lake for both streaming and batch data. We are also responsible for generating the foundational data. The systems we provide include Scribe, Kafka, Hive, Presto, Spark, Flink, Pinot, and others. The team is actively involved in Open Source, and we are planning to increase our engagement over time.

About the Role We are seeking a skilled engineer with exceptional DevOps skills to join our team. Responsibilities include automating and scaling Big Data and Analytics technology stacks on Cloud infrastructure, building CI/CD pipelines, setting up monitoring and alerting for production infrastructure, and keeping our technology stacks up to date.

For California Only - The estimated annual salary for this position is between $186,000 - $340,000 annually. Compensation packages are based on factors unique to each candidate, including but not limited to skill set, certifications, and specific geographical location. This role is eligible for health insurance, equity awards, life insurance, disability benefits, parental leave, wellness benefits, and paid time off.

What you’ll be doing:

Develop best practices around cloud infrastructure provisioning, disaster recovery, and guiding developers on the adoption

Scale Big Data and distributed systems

Collaborate on system architecture with developers for optimal scaling, resource utilization, fault tolerance, reliability, and availability

Conduct low-level systems debugging, performance measurement & optimization on large production clusters and low-latency services

Create scripts and automation that can react quickly to infrastructure issues and take corrective actions

Participate in architecture discussions, influence product roadmap, and take ownership and responsibility over new projects

Collaborate and communicate with a geographically distributed team

We’re excited if you have:

Bachelor’s degree, or equivalent work experience

8+ years of experience in DevOps or Site Reliability Engineering

Experience with Cloud infrastructure such as Amazon AWS, Google Cloud Platform (GCP), Microsoft Azure, or other Public Cloud platforms. GCP is preferred.

Experience with at least 3 of the technologies/tools mentioned here: Big Data / Hadoop, Kafka, Spark, Airflow, Presto, Druid, Opensearch, HA Proxy, or Hive

Experience with Kubernetes and Docker

#J-18808-Ljbffr