Logo
Granica

Software Engineer, AI Data Platform (Mountain View)

Granica, Mountain View, California, United States, 94039

Save Job

Granica is redefining how enterprises prepare and optimize data at the most fundamental layer of the AI stackwhere raw information becomes usable intelligence. Our technology operates deep in the data infrastructure layer, making data efficient, secure, and ready for scale. We eliminate the hidden inefficiencies in modern data platformsslashing storage and compute costs, accelerating pipelines, and boosting platform efficiency. The result: 60%+ lower storage costs, up to 60% lower compute spend, 3 faster data processing, and 20% overall efficiency gains.

Why It Matters Massive data should fuel innovation, not drain budgets. We remove the bottlenecks holding AI and analytics backmaking data lighter, faster, and smarter so teams can ship breakthroughs, not babysit storage and compute bills.

Who We Are World-renowned researchers in compression, information theory, and data systems Elite engineers from Google, Pure Storage, Cohesity, and top cloud teams Enterprise sellers who turn ROI into sevenfigure wins.

Powered by World-Class Investors & Customers $65M+ raised from NEA, Bain Capital, A* Capital, and operators behind Okta, Eventbrite, Tesla, and Databricks. Our platform already processes hundreds of petabytes for industry leaders

Our Mission:

Were building the default data substrate for AI, and a generational company built to endure. Smarter Infrastructure for the AI Era: We make data efficient, safe, and ready for scalethink smarter, more foundational infrastructure for the AI era. Our technology integrates directly with modern data stacks like Snowflake, Databricks, and S3-based data lakes, enabling: 60%+ reduction in storage costs and up to 60% lower compute spend 3x faster data processing 20% platform efficiency gains

Trusted by Industry Leaders Enterprise leaders globally already rely on Granica to cut costs, boost performance, and unlock more value from their existing data platforms.

A Deep Tech Approach to AI Were unlocking the layers

beneath

platforms like Snowflake and Databricks, making them faster, cheaper, and more AI-native. We combine advanced research with practical productization, powered by a dual-track strategy: Research:

Led by Chief Scientist Andrea Montanari (Stanford Professor), we publish 12 top-tier papers per quarter. Product:

Actively processing 100+ PBs today and targeting Exabyte scale by Q4 2025.

Backed by the Best Weve raised $60M+ from NEA, Bain Capital, A Capital, and operators behind Okta, Eventbrite, Tesla, and Databricks.

Our Mission To

convert entropy into intelligence , so every builderhuman or AIcan make the impossible real. Were building the

default data substrate for AI , and a generational company built to endure beyond any single product cycle.

WHAT YOULL DO This is a deep systems role for someone who lives and breathes distributed infrastructure, understands how data moves at scale, and wants to build the next-generation

AI data platform

from the ground up. Own the ACID backbone.

Design and harden transactional layers and metadata services so that petabyte-scale tables can time-travel in microseconds and schema evolution becomes a non-event. Turn metadata into rocket fuel.

Build compaction, caching, and pruning services that keep millions of file pointers within 50 ms from lookup to plan. Squeeze more signal per byte.

Optimize data layoutsfrom column ordering to dictionary and bit-packing, bloom filters, and zone-map indexesto cut scan I/O by 10 on real-world workloads. Ship adaptive indexing with research.

Co-invent machine-driven indexes that learn access patterns and automatically re-partition nightlyno more manual analyze table ever again. Scale the engine, not the babysitting.

Write Spark, Flink, or batch pipelines that autoscale across S3, GCS, and ADLS; expose observability hooks; and survive chaos drills without triggering a pager storm. Code for longevity.

Write clean, test-soaked Java, Scala, Go, or C++. Document key invariants so future teams can extend the systeminstead of rewriting it. Measure success in human latency.

If analysts see their dashboards refresh in blink-level time, youve won. Publish your breakthrough and mentor the next engineer to raise the bar again.

WHAT WERE LOOKING FOR Youve built systems where performance, resilience, and clarity of design all matter. You thrive at the intersection of infrastructure engineering and applied research, and care deeply about both how something works and how well it works at scale.

Core Skills Distributed Systems and Storage Fundamentals

consistency, replication, sharding, durability, transactions. Columnar Storage Optimization

deep knowledge of Parquet or similar formats (column ordering, compression, zone maps). Metadata and Indexing Systems

experience building metadata-driven services, compaction, caching, and adaptive indexing. Distributed Compute at Scale

production-grade Spark/Flink or equivalent pipeline development across S3, GCS, or ADLS. Programming for Scale and Longevity

strong coding in Java, Scala, Go, or C++, with clean testing and documentation practices. Resilient Systems and Observability

youve built systems that survive chaos drills and expose the right metrics.

Desired Skills Exposure to

open table formats

such as Apache Iceberg, Delta Lake, or Hudi. Experience with

catalog services, query planning, or compaction frameworks . OSS contributions or published work in data infrastructure or distributed systems.

WHY JOIN GRANICA If youve helped build the modern data stack at a large companyDatabricks, Snowflake, Confluent, or similaryou already know how critical lakehouse infrastructure is to AI and analytics at scale. At Granica, youll take that knowledge and apply it where it matters mostat the most fundamental layer in the data ecosystem. Own the product, not just the feature.

At Granica, you wont be optimizing edge cases or maintaining legacy systems. Youll architect and build foundational components that define how enterprises manage and optimize data for AI. Move faster, go deeper.

No multi-month review cycles or layers of abstractionjust high-agency engineering work where great ideas ship weekly. Youll work directly with the founding team, engage closely with design partners, and see your impact hit production fast. Work on hard, meaningful problems.

From transaction layer design in Delta and Iceberg, to petabyte-scale compaction and schema evolution, to adaptive indexing and cost-aware query planningthis is deep systems engineering at scale. Join a team of expert builders.

Our engineers have designed the core internals of cloud-scale data systems, and we maintain a culture of peer-driven learning, hands-on prototyping, and technical storytelling. Core Differentiation:

Were

focused on unlocking a deeper layer of AI infrastructure. By optimizing the way data is stored, processed, and retrieved, we make platforms like Snowflake and Databricks faster, more cost-efficient, and more AI-native. Our work sits at the most fundamental layer of the AI stack: where raw data becomes usable intelligence. Be part of something earlywithout the chaos.

Granica has already secured $65M+ from NEA, Bain Capital Ventures, A* Capital, and legendary operators from Okta, Tesla, and Databricks. Grow with the company.

Youll have the chance to grow into a technical leadership role, mentor future hires, and shape both the engineering culture and product direction as we scale.

COMPENSATION & BENEFITS Competitive salary and meaningful equity Unlimited PTO + quarterly recharge days Premium health, vision, and dental Team offsites, deep tech talks, and learning stipends Help build the

foundational infrastructure for the AI era

Granica is an equal opportunity employer.

We celebrate diversity and are committed to creating an inclusive environment for all employees.