Cast AI
Company Overview
Cast AI is the leading Application Performance Automation (APA) platform that helps customers cut cloud costs, improve performance, and boost productivity automatically. Built originally for Kubernetes, it delivers real‑time autonomous optimization across any cloud environment, continuously analyzing workloads, rightsizing resources, and rebalancing clusters without manual intervention. Headquartered in Miami, Florida, the company operates in more than 32 countries and serves over 2,100 companies including BMW, Akamai, Hugging Face, and NielsenIQ.
About the Role AI Enabler
– Helps customers deploy and manage LLMs in their Kubernetes clusters and optimizes workloads by providing cost visibility and intelligent routing for LLM requests to the most cost‑effective compute resources (e.g., Grok, self‑hosted LLAMA models).
Tools We Use
Languages: Go (primary), Python (secondary)
Cloud & Orchestration: Kubernetes, AWS, GCP, Azure
Databases & Storage: PostgreSQL, Cloud Object Storage
Messaging & APIs: GCP Pub/Sub, gRPC, REST
Observability: Prometheus, Grafana, Loki, Tempo
CI/CD & GitOps: GitLab CI, ArgoCD
Requirements
Strong software engineering skills, experience in distributed systems and back‑end development (preferably Go, but transition willingness acceptable)
Debugging, optimization, and performance tuning expertise
Hands‑on experience with AWS, GCP, Azure, and Kubernetes
Experience with CI/CD and DevOps practices
Strong written and verbal English
Ability to work independently and collaboratively in a team
Startup mindset: adaptable, proactive, comfortable with ambiguity
Proactive, problem‑solving attitude with a "yes we can" mentality
Benefits
Competitive salary (€6,500 - €9,000 gross, depending on experience)
Flexible, remote‑first global environment
Equity options
Private health insurance
Fast‑paced workflow with feature projects completed in 1–4 weeks
10% of work time dedicated to personal projects or self‑improvement
Learning budget for professional and personal development, including conferences and courses
Annual hackathon and team‑building events
Equipment budget and extra days off for work‑life balance
Details
Seniority level: Mid‑Senior
Employment type: Full‑time
Job function: Information Technology
Industries: Software Development
#J-18808-Ljbffr
About the Role AI Enabler
– Helps customers deploy and manage LLMs in their Kubernetes clusters and optimizes workloads by providing cost visibility and intelligent routing for LLM requests to the most cost‑effective compute resources (e.g., Grok, self‑hosted LLAMA models).
Tools We Use
Languages: Go (primary), Python (secondary)
Cloud & Orchestration: Kubernetes, AWS, GCP, Azure
Databases & Storage: PostgreSQL, Cloud Object Storage
Messaging & APIs: GCP Pub/Sub, gRPC, REST
Observability: Prometheus, Grafana, Loki, Tempo
CI/CD & GitOps: GitLab CI, ArgoCD
Requirements
Strong software engineering skills, experience in distributed systems and back‑end development (preferably Go, but transition willingness acceptable)
Debugging, optimization, and performance tuning expertise
Hands‑on experience with AWS, GCP, Azure, and Kubernetes
Experience with CI/CD and DevOps practices
Strong written and verbal English
Ability to work independently and collaboratively in a team
Startup mindset: adaptable, proactive, comfortable with ambiguity
Proactive, problem‑solving attitude with a "yes we can" mentality
Benefits
Competitive salary (€6,500 - €9,000 gross, depending on experience)
Flexible, remote‑first global environment
Equity options
Private health insurance
Fast‑paced workflow with feature projects completed in 1–4 weeks
10% of work time dedicated to personal projects or self‑improvement
Learning budget for professional and personal development, including conferences and courses
Annual hackathon and team‑building events
Equipment budget and extra days off for work‑life balance
Details
Seniority level: Mid‑Senior
Employment type: Full‑time
Job function: Information Technology
Industries: Software Development
#J-18808-Ljbffr