Onto Innovation
Overview
Onto Innovation is a worldwide leader in the design, development, manufacture and support of defect inspection, advanced packaging lithography, process control metrology, and data analysis systems and software used by semiconductor device manufacturers worldwide. Onto Innovation provides a full-fab solution through its families of proprietary products that provide critical yield-enhancing information and real time process control responses, enabling microelectronic device manufacturers to drive down the costs and time to market of their products. The Company’s expanding portfolio of equipment and software solutions is used in both the wafer processing and final manufacturing of ICs, and in adjacent markets such as FPD, and LED manufacturing. Responsibilities
Prototype AI assistants & agents for field workflows: guided recipe setup, log triage, playbook lookups, parts/alarms advice, and fleet-wide health checks. Build retrieval systems (RAG): ingest manuals, specs, ticket notes, recipes, logs, and best-practice docs; design chunking, embeddings, and indexing; tune prompts and retrieval for accuracy/latency. Connect AI to our tools and data: stand up MCP servers (Model Context Protocol) and other connectors to safely expose internal systems (document stores, MES, issue trackers, telemetry APIs) to LLMs. Fine-tune or adapt models (e.g., LoRA/QLoRA) for domain terms, error codes, and tool-specific intents when retrieval alone isn’t enough. Evaluate and harden: set up offline & online evals for groundedness/relevance; add guardrails, observability, and traceability; write runbooks. Ship small apps: package prototypes behind simple APIs or lightweight UIs that field engineers can use (web chat, Slack/Teams bots, or CLI). Data plumbing: parse messy PDFs/images/CSVs; normalize schemas for recipes, events, alarms, SPC/trace data. Computer Vision: understanding, defect detection, segmentation, or SEM/optical imaging. Work like an engineer: write readable Python/TypeScript, tests, and docs; use Git; participate in code reviews; iterate fast with the AI lead and domain SMEs. Minimum qualifications
BS in CS/EE/CE/ME (or equivalent experience). Python
proficiency (data wrangling, APIs, packaging); comfort on
Linux
and with Git. Built at least one LLM app using a framework such as
LangChain, LlamaIndex, or Semantic Kernel . Hands-on with
vector search
(e.g., FAISS/Weaviate/Milvus) and embeddings; understands chunking, metadata, and hybrid search basics. Familiarity with
RAG
and prompt engineering; can measure quality (groundedness/relevance) and reduce hallucinations. Basic backend skills (REST/JSON, auth, environment secrets); experience containerizing with
Docker . Comfortable reading technical manuals/logs and collaborating with non-software teammates. Nice to have
Worked with
agent frameworks
(LangGraph, AutoGen, CrewAI) or implemented tool-calling/plan-execute loops. Built or configured
MCP servers
to connect LLMs to internal data/tools. Experience parsing complex docs (e.g.,
Unstructured ,
GROBID ) and handling images/figures from manuals. Exposure to
semiconductor equipment
or factory systems (SECS/GEM, EDA/Interface A, MES, SPC); familiarity with
KLA/AMAT/TEL/ASML
tool ecosystems. Time-series and log analysis (Pandas, SQL, TimescaleDB/InfluxDB), wafer map/vision background, or simple CV. Model adaptation experience ( LoRA/QLoRA , PEFT) and experiment tracking (MLflow/W&B). LLM observability/evals (Ragas, TruLens, LangSmith), basic security/PII handling, and role-based access. Cloud familiarity (AWS/Azure/GCP) and lightweight front-ends (React/Next.js) for internal tools. Prior work on
fleet-level
dashboards/analytics or recipe/parameter management. What success looks like (first 90 days)
Ship a
search+chat knowledge assistant
over our internal docs with clear eval dashboards for faithfulness/relevance. Stand up at least one
MCP connector
to an internal source (e.g., SharePoint/Confluence or log store) and demo safe tool calls. Deliver a focused
POC : e.g., an agent that reads recent alarms & logs to suggest next steps, or a fleet health summary with links to playbooks. Document everything (design notes, runbooks, and “how to” guides) and gather field feedback for iteration. How we work
Pragmatic, security-minded, iterate-in-the-open with our engineers. We value curiosity, clear writing, and the grit to trace weird edge cases in logs and manuals. Apply
Send your resume/GitHub/portfolio and a short note about an LLM or agent project you’ve built (what made it work, what you measured, and what you’d improve). Qualifications
see above Onto Innovation Inc. offers competitive salaries and a generous benefits package, including health/dental/vision/life/disability, PTO, 401K plan with employer match, and an Employee Stock Purchase Program (ESPP) along with health & wellness initiatives. We provide a collaborative working environment along with resources, and state-of-the-art tools & equipment to promote success; and a welcoming, inclusive corporate culture where individuals are recognized for their contributions. Onto Innovation Inc. is an Equal Opportunity Employer and all qualified applicants will receive consideration for employment without regard to race, color, religion, gender, sexual orientation, national origin, genetic information, age, disability, veteran status, or any other legally protected basis. For positions requiring access to technical data, Onto Innovation Inc., Inc. may have to obtain export licensing approval from the U.S. Department of Commerce - Bureau of Industry and Security and/or the U.S. Department of State - Directorate of Defense Trade Controls. As such, applicants for this position – except US Citizens, US Permanent Residents, and protected individuals as defined by 8 U.S.C. 1324b(a)(3) – may have to go through an export licensing review process.
#J-18808-Ljbffr
Onto Innovation is a worldwide leader in the design, development, manufacture and support of defect inspection, advanced packaging lithography, process control metrology, and data analysis systems and software used by semiconductor device manufacturers worldwide. Onto Innovation provides a full-fab solution through its families of proprietary products that provide critical yield-enhancing information and real time process control responses, enabling microelectronic device manufacturers to drive down the costs and time to market of their products. The Company’s expanding portfolio of equipment and software solutions is used in both the wafer processing and final manufacturing of ICs, and in adjacent markets such as FPD, and LED manufacturing. Responsibilities
Prototype AI assistants & agents for field workflows: guided recipe setup, log triage, playbook lookups, parts/alarms advice, and fleet-wide health checks. Build retrieval systems (RAG): ingest manuals, specs, ticket notes, recipes, logs, and best-practice docs; design chunking, embeddings, and indexing; tune prompts and retrieval for accuracy/latency. Connect AI to our tools and data: stand up MCP servers (Model Context Protocol) and other connectors to safely expose internal systems (document stores, MES, issue trackers, telemetry APIs) to LLMs. Fine-tune or adapt models (e.g., LoRA/QLoRA) for domain terms, error codes, and tool-specific intents when retrieval alone isn’t enough. Evaluate and harden: set up offline & online evals for groundedness/relevance; add guardrails, observability, and traceability; write runbooks. Ship small apps: package prototypes behind simple APIs or lightweight UIs that field engineers can use (web chat, Slack/Teams bots, or CLI). Data plumbing: parse messy PDFs/images/CSVs; normalize schemas for recipes, events, alarms, SPC/trace data. Computer Vision: understanding, defect detection, segmentation, or SEM/optical imaging. Work like an engineer: write readable Python/TypeScript, tests, and docs; use Git; participate in code reviews; iterate fast with the AI lead and domain SMEs. Minimum qualifications
BS in CS/EE/CE/ME (or equivalent experience). Python
proficiency (data wrangling, APIs, packaging); comfort on
Linux
and with Git. Built at least one LLM app using a framework such as
LangChain, LlamaIndex, or Semantic Kernel . Hands-on with
vector search
(e.g., FAISS/Weaviate/Milvus) and embeddings; understands chunking, metadata, and hybrid search basics. Familiarity with
RAG
and prompt engineering; can measure quality (groundedness/relevance) and reduce hallucinations. Basic backend skills (REST/JSON, auth, environment secrets); experience containerizing with
Docker . Comfortable reading technical manuals/logs and collaborating with non-software teammates. Nice to have
Worked with
agent frameworks
(LangGraph, AutoGen, CrewAI) or implemented tool-calling/plan-execute loops. Built or configured
MCP servers
to connect LLMs to internal data/tools. Experience parsing complex docs (e.g.,
Unstructured ,
GROBID ) and handling images/figures from manuals. Exposure to
semiconductor equipment
or factory systems (SECS/GEM, EDA/Interface A, MES, SPC); familiarity with
KLA/AMAT/TEL/ASML
tool ecosystems. Time-series and log analysis (Pandas, SQL, TimescaleDB/InfluxDB), wafer map/vision background, or simple CV. Model adaptation experience ( LoRA/QLoRA , PEFT) and experiment tracking (MLflow/W&B). LLM observability/evals (Ragas, TruLens, LangSmith), basic security/PII handling, and role-based access. Cloud familiarity (AWS/Azure/GCP) and lightweight front-ends (React/Next.js) for internal tools. Prior work on
fleet-level
dashboards/analytics or recipe/parameter management. What success looks like (first 90 days)
Ship a
search+chat knowledge assistant
over our internal docs with clear eval dashboards for faithfulness/relevance. Stand up at least one
MCP connector
to an internal source (e.g., SharePoint/Confluence or log store) and demo safe tool calls. Deliver a focused
POC : e.g., an agent that reads recent alarms & logs to suggest next steps, or a fleet health summary with links to playbooks. Document everything (design notes, runbooks, and “how to” guides) and gather field feedback for iteration. How we work
Pragmatic, security-minded, iterate-in-the-open with our engineers. We value curiosity, clear writing, and the grit to trace weird edge cases in logs and manuals. Apply
Send your resume/GitHub/portfolio and a short note about an LLM or agent project you’ve built (what made it work, what you measured, and what you’d improve). Qualifications
see above Onto Innovation Inc. offers competitive salaries and a generous benefits package, including health/dental/vision/life/disability, PTO, 401K plan with employer match, and an Employee Stock Purchase Program (ESPP) along with health & wellness initiatives. We provide a collaborative working environment along with resources, and state-of-the-art tools & equipment to promote success; and a welcoming, inclusive corporate culture where individuals are recognized for their contributions. Onto Innovation Inc. is an Equal Opportunity Employer and all qualified applicants will receive consideration for employment without regard to race, color, religion, gender, sexual orientation, national origin, genetic information, age, disability, veteran status, or any other legally protected basis. For positions requiring access to technical data, Onto Innovation Inc., Inc. may have to obtain export licensing approval from the U.S. Department of Commerce - Bureau of Industry and Security and/or the U.S. Department of State - Directorate of Defense Trade Controls. As such, applicants for this position – except US Citizens, US Permanent Residents, and protected individuals as defined by 8 U.S.C. 1324b(a)(3) – may have to go through an export licensing review process.
#J-18808-Ljbffr