Hawaii Staffing
Senior Principal Software Engineer- OCI AI Services
Hawaii Staffing, Honolulu, Hawaii, United States, 96814
Consulting Member of Technical Staff
At Oracle Cloud Infrastructure (OCI), we are redefining the future of computing for enterprisesbuilding cloud-native systems from the ground up, powered by a global team of visionary engineers, scientists, and creators. We combine the agility of a startup with the scale, security, and reach of Oracle's enterprise-grade platforms. Our Generative AI Service team is pioneering the development of infrastructure and services that harness the transformative power of Large Language Models (LLMs) and Agentic AI systems. Our mission is to build world-class, scalable platforms that enable customers to deploy intelligent agents and applications, deeply integrated with OCI's robust cloud ecosystem. Role Summary: As a Consulting Member of Technical Staff (IC5), you will play a pivotal role in designing, building, and optimizing LLM infrastructure, agent execution runtimes, and next-generation developer platforms. You'll collaborate closely with applied scientists and ML engineers to bring agentic workflows into real-world deploymentsat scale. This is a hands-on technical leadership role, ideal for someone deeply rooted in distributed systems and low-level computer science. Minimum Qualifications
BS in Computer Science or equivalent experience. 10+ years of experience in production-grade distributed systems and cloud-native software engineering. Proficiency in Go, Java, Python, or C++. Expertise in high-performance computing and ML model serving infrastructure. Deep understanding of container orchestration and CI/CD pipelines. Strong communication skills and experience mentoring across teams. Preferred Qualifications
MS or PhD in Computer Science, particularly in Systems, ML Infrastructure, or Compilers. Experience with LLM serving frameworks like vLLM, FasterTransformer, DeepSpeed, or Triton. Familiarity with agent-based systems. Contributions to LLM-native developer tools and compiler IRs. Experience with vector databases, tool APIs, and event-driven workflows. Foundation in OS internals, compiler pipelines, and systems programming. Proven ability to lead large-scale architecture efforts. Why Join Us: Be at the frontier of generative AI and agent-based software interaction. Work on mission-critical projects impacting Oracle's AI strategy. Collaborate with a globally distributed team of leading engineers and researchers. Enjoy the agility of a fast-moving team with enterprise-level resources. Responsibilities: Architect and build high-throughput, low-latency serving systems for LLM inference and agent orchestration. Design agent-native runtime environments that support dynamic planning, tool calling, memory, and long-running context. Integrate foundational AI components with OCI's compute and networking layers. Partner with ML research to optimize model training, fine-tuning, and inference performance on GPU clusters. Own critical paths of software delivery from architectural review through implementation and post-deployment resilience. Contribute to OCI's developer-facing agent framework. Tackle deep systems-level challenges, drawing on knowledge in operating systems, compiler design, and cloud primitives. Disclaimer: Certain US customer or client-facing roles may be required to comply with applicable requirements, such as immunization and occupational health mandates. Range and benefit information provided in this posting are specific to the stated locations only. US: Hiring Range in USD from: $96,800 to $251,600 per annum. May be eligible for bonus, equity, and compensation deferral.
At Oracle Cloud Infrastructure (OCI), we are redefining the future of computing for enterprisesbuilding cloud-native systems from the ground up, powered by a global team of visionary engineers, scientists, and creators. We combine the agility of a startup with the scale, security, and reach of Oracle's enterprise-grade platforms. Our Generative AI Service team is pioneering the development of infrastructure and services that harness the transformative power of Large Language Models (LLMs) and Agentic AI systems. Our mission is to build world-class, scalable platforms that enable customers to deploy intelligent agents and applications, deeply integrated with OCI's robust cloud ecosystem. Role Summary: As a Consulting Member of Technical Staff (IC5), you will play a pivotal role in designing, building, and optimizing LLM infrastructure, agent execution runtimes, and next-generation developer platforms. You'll collaborate closely with applied scientists and ML engineers to bring agentic workflows into real-world deploymentsat scale. This is a hands-on technical leadership role, ideal for someone deeply rooted in distributed systems and low-level computer science. Minimum Qualifications
BS in Computer Science or equivalent experience. 10+ years of experience in production-grade distributed systems and cloud-native software engineering. Proficiency in Go, Java, Python, or C++. Expertise in high-performance computing and ML model serving infrastructure. Deep understanding of container orchestration and CI/CD pipelines. Strong communication skills and experience mentoring across teams. Preferred Qualifications
MS or PhD in Computer Science, particularly in Systems, ML Infrastructure, or Compilers. Experience with LLM serving frameworks like vLLM, FasterTransformer, DeepSpeed, or Triton. Familiarity with agent-based systems. Contributions to LLM-native developer tools and compiler IRs. Experience with vector databases, tool APIs, and event-driven workflows. Foundation in OS internals, compiler pipelines, and systems programming. Proven ability to lead large-scale architecture efforts. Why Join Us: Be at the frontier of generative AI and agent-based software interaction. Work on mission-critical projects impacting Oracle's AI strategy. Collaborate with a globally distributed team of leading engineers and researchers. Enjoy the agility of a fast-moving team with enterprise-level resources. Responsibilities: Architect and build high-throughput, low-latency serving systems for LLM inference and agent orchestration. Design agent-native runtime environments that support dynamic planning, tool calling, memory, and long-running context. Integrate foundational AI components with OCI's compute and networking layers. Partner with ML research to optimize model training, fine-tuning, and inference performance on GPU clusters. Own critical paths of software delivery from architectural review through implementation and post-deployment resilience. Contribute to OCI's developer-facing agent framework. Tackle deep systems-level challenges, drawing on knowledge in operating systems, compiler design, and cloud primitives. Disclaimer: Certain US customer or client-facing roles may be required to comply with applicable requirements, such as immunization and occupational health mandates. Range and benefit information provided in this posting are specific to the stated locations only. US: Hiring Range in USD from: $96,800 to $251,600 per annum. May be eligible for bonus, equity, and compensation deferral.