Logo
Outspeed

Member of Technical Staff - Outspeed

Outspeed, Little Rock, Arkansas, United States

Save Job

About Us Outspeed is solving one of the biggest challenges with current AI systems - latency. We are building the infra for real-time AI for applications in gaming, AR/VR, robotics, and more.

Outspeed is led by an experienced team of researchers and engineers with collective experience from MIT, Google, and Microsoft. Our team is based in San Francisco and values empathy, deep technical knowledge, and autonomy.

Janak earned his bachelor's and master's degrees from MIT and built grid-scale AI algorithms and infra at Autogrid (acq. Schneider). Sahil is a published researcher who led AI infra efforts at Google and Microsoft.

Role Description As an early Member of Technical Staff you’ll dive into every layer of Outspeed’s real-time AI platform—core inference engines, orchestration services, developer APIs, and customer-facing tools. You’ll alternate between rapid prototyping and hardening production systems, shipping code that is immediately exercised by teams around the world.

Typical weeks might include:

Owning end-to-end feature work—from RFC and design review to infrastructure-as-code, monitoring, and post-launch iteration.

Optimizing GPU inference pipelines, extending our TypeScript/React console, and designing a new audio-streaming protocols.

Pairing with customers to debug latency spikes, then upstreaming the fix into our orchestration layer.

Influencing engineering culture: instituting best practices, mentoring newer hires, and shaping our technical roadmap alongside the founders.

This role is perfect for an engineer who enjoys breadth, thrives on context-switching, and wants their fingerprints on everything we build.

Benefits

Competitive salary + Equity

Health, dental, and vision insurance.

Bonuses based on performance

We are a company founded by immigrants, and we are committed to providing support to immigrant workers throughout their journey.

Requirements

2+ years

of professional software engineering, with

deep proficiency in at least one systems language (Go/Rust/C++) and one high-level language (Python/TypeScript).

Proven track record shipping

production-grade distributed systems —you’ve designed, implemented, and operated services that run at scale and stay up.

Hands-on experience with

end-to-end ML workflows : data pipelines, model training/tuning, packaging, and low-latency serving (PyTorch + CUDA or similar).

Strong product sense and the ability to translate ambiguous problems into well-scoped engineering work.

Excellent written & verbal communication; comfortable leading design docs and code reviews in a fast-moving, asynchronous environment.

Nice-to-haves

Expertise running containerized workloads with

Docker, Kubernetes, or Nomad , plus IaC with

Terraform or Pulumi .

Knowledge of public-cloud primitives (SDN, block storage, secrets/identity, spot fleet management) and how to squeeze every dollar out of them.

Familiarity with modern

inference tooling

(e.g., vLLM, Ray, Triton, llama.cpp) and a solid mental model of GPU/TPU performance bottlenecks.

A history of thriving in early-stage startups: rapid iteration, “build → measure → learn” loops, and wearing multiple hats at once.

#J-18808-Ljbffr