Logo
Weights & Biases

Principal Product Manager, W&B Inference - Weights & Biases

Weights & Biases, Sunnyvale, California, United States, 94087

Save Job

Principal Product Manager, W&B Inference - Weights & Biases Join to apply for the Principal Product Manager, W&B Inference role at Weights & Biases.

CoreWeave, the AI Hyperscaler™, acquired Weights & Biases to create the most powerful end‑to‑end platform to develop, deploy, and iterate AI faster. Since 2017, CoreWeave has operated a growing footprint of data centers covering every region of the US and across Europe, and was ranked as one of the TIME100 most influential companies of 2024. By bringing together CoreWeave’s industry‑leading cloud infrastructure with the best‑in‑class tools AI practitioners know and love from Weights & Biases, we’re setting a new standard for how AI is built, trained, and scaled.

The integration of our teams and technologies is accelerating our shared mission: to empower developers with the tools and infrastructure they need to push the boundaries of what AI can do. From experiment tracking and model optimization to high‑performance training clusters, agent building, and inference at scale, we’re combining forces to serve the full AI lifecycle — all in one seamless platform.

Weights & Biases has long been trusted by over 1,500 organizations — including AstraZeneca, Canva, Cohere, OpenAI, Meta, Snowflake, Square, Toyota, and Wayve — to build better models, AI agents and applications. Now, as part of CoreWeave, that impact is amplified across a broader ecosystem of AI innovators, researchers, and enterprises.

As we unite under one vision, we’re looking for bold thinkers and agile builders who are excited to shape the future of AI alongside us. If you're passionate about solving complex problems at the intersection of software, hardware, and AI, there's never been a more exciting time to join our team.

What You’ll Do As a Principal Product Manager for the Weights & Biases Inference Service, you will be responsible for defining, driving, and scaling one of the most critical components of our platform. You’ll own the vision, roadmap, and execution for the inference service end‑to‑end, ensuring it evolves to meet the performance, reliability, and usability needs of our customers. In this role, you will lead mission‑critical initiatives that strengthen and extend the core infrastructure powering W&B’s machine‑learning workflows.

About The Role

Own the execution and evolution of the W&B Inference Service, delivering solutions that directly support the product vision and long‑term platform strategy.

Lead cross‑team initiatives end‑to‑end, coordinating engineering, product, security, operations, and go‑to‑market stakeholders to ensure aligned priorities and seamless delivery across interdependent systems.

Prioritize with intention, making informed trade‑offs among performance, reliability, compliance, cost, and development velocity to ensure the inference service scales to meet customer and platform demands.

Elevate developer and practitioner experiences by improving the operability, observability, and usability of the inference service and the tooling that surrounds it.

Own execution from requirements through launch, defining success metrics, gathering customer and system insights, and ensuring every stage of development is anchored in measurable outcomes.

Who You Are

Experience: A seasoned product manager with 7+ years working on high‑scale platform or infrastructure products, with direct experience in model serving, inference systems, real‑time APIs, or distributed compute services.

Inference systems versatility: You’ve worked across domains that commonly intersect with inference—including autoscaling, observability, GPU/accelerator utilization, routing/orchestration, developer tooling, IAM, and storage—and can reason about how changes ripple through a real‑time serving stack.

Deep technical fluency: You can engage engineers on service architectures, performance bottlenecks, deployment topologies, model packaging formats, request/response patterns, and reliability trade‑offs that impact low‑latency inference. You’re comfortable interpreting architecture diagrams and discussing how design decisions influence throughput, cost, and SLAs.

Cross‑functional leadership: Adept at coordinating teams across inference runtime, infrastructure, security, operations, and go‑to‑market, ensuring alignment on priorities that improve the performance, reliability, and usability of the inference service.

Customer empathy for practitioners: You understand the workflows of ML practitioners running production models and the needs of internal developers building on top of the inference platform. You’re motivated by uncovering friction in their serving pipelines and translating those insights into meaningful improvements.

Execution mindset: You excel in ambiguous, fast‑moving environments. You bring clarity to competing priorities, make thoughtful trade‑offs among latency, reliability, cost, and velocity, and consistently drive inference‑focused initiatives from concept to launch.

Preferred

Direct experience as a PM for an inference or model‑serving service, ideally involving real‑time, low‑latency, or high‑throughput workloads. Experience with frameworks like TensorFlow, PyTorch, or model‑serialization formats is a plus.

Background in adjacent platform domains such as identity & access management, billing and metering workflows, observability, or data infrastructure—especially where they intersect with running models in production.

Strong familiarity with cloud infrastructure (AWS, GCP, Azure), container orchestration, autoscaling, and deployment automation tools used to operate distributed inference systems.

Exposure to W&B or similar MLOps tools, especially experiment tracking, model management, or deployment workflows.

Wondering if you’re a good fit?

You love to build frictionless products for developers

You’re curious about AI and MLOps tooling

You’re an expert in building inference systems that scale for production workloads

Why Us? We work hard, have fun, and move fast! We’re in an exciting stage of hyper‑growth that you will not want to miss out on. We’re not afraid of a little chaos, and we’re constantly learning. Our team cares deeply about how we build our product and how we work together, which is represented through our core values: Be Curious at Your Core, Act Like an Owner, Empower Employees, Deliver Best‑in‑Class Client Experiences, Achieve More Together.

What We Offer The base salary range for this role is $206,000 to $303,000. The starting salary will be determined based on job‑related knowledge, skills, experience, and market location. We strive for both market alignment and internal equity when determining compensation. In addition to base salary, our total rewards package includes a discretionary bonus, equity awards, and a comprehensive benefits program (all based on eligibility).

Medical, dental, and vision insurance – 100% paid for by CoreWeave

Company‑paid Life Insurance

Voluntary supplemental life insurance

Short and long‑term disability insurance

Flexible Spending Account

Health Savings Account

Tuition Reimbursement

Ability to Participate in Employee Stock Purchase Program (ESPP)

Mental Wellness Benefits through Spring Health

Family‑Forming support provided by Carrot

Paid Parental Leave

Flexible, full‑service childcare support with Kinside

401(k) with a generous employer match

Flexible PTO

Catered lunch each day in our office and data center locations

A casual work environment

A work culture focused on innovative disruption

Our Workplace While we prioritize a hybrid work environment, remote work may be considered for candidates located more than 30 miles from an office, based on role requirements for specialized skill sets. New hires will be invited to attend onboarding at one of our hubs within their first month. Teams also gather quarterly to support collaboration.

California Consumer Privacy Act – California applicants only.

Equal Employment Opportunity CoreWeave is an equal opportunity employer, committed to fostering an inclusive and supportive workplace. All qualified applicants and candidates will receive consideration for employment without regard to race, color, religion, sex, disability, age, sexual orientation, gender identity, national origin, veteran status, or genetic information. As part of this commitment and consistent with the Americans with Disabilities Act (ADA), CoreWeave will ensure that qualified applicants and candidates with disabilities are provided reasonable accommodations for the hiring process, unless such accommodation would cause an undue hardship. If reasonable accommodation is needed, please contact: careers@coreweave.com.

Export Control Compliance This position requires access to export‑controlled information. To conform to U.S. Government export regulations applicable to that information, applicant must either be:

1157, or (iv) asylee under 8 U.S.C. 1158, a U.S. person, a U.S. citizen or national, a U.S. lawful permanent resident (green card holder), or refugee under 8 U.S.C. 1157.

Eligible to access the export‑controlled information without a required export authorization, or

Eligible and reasonably likely to obtain the required export authorization from the applicable U.S. government agency. CoreWeave may, for legitimate business reasons, decline to pursue any export licensing process.

Seniority level: Mid‑Senior level

Employment type: Full‑time

Job function: Product Management and Marketing; Industries: Software Development

#J-18808-Ljbffr