The Hartford
Sr Staff Data Engineer - Hybrid
The Hartford, Charlotte, North Carolina, United States, 28245
Join to apply for the
Sr Staff Data Engineer - Hybrid
role at
The Hartford
Sr Staff AI Data Engineer is responsible for implementing AI data pipelines that bring together structured, semi‑structured and unstructured data to support AI and Agentic solutions. This includes pre‑processing with extraction, chunking, embedding and grounding strategies to get the data ready.
This role will have a Hybrid work schedule, with the expectation of working in an office location (Hartford, CT; Chicago, IL; Columbus, OH; and Charlotte, NC) 3 days a week (Tuesday through Thursday).
Responsibilities
AI Data Engineering lead responsible for implementing AI data pipelines that bring together structured, semi‑structured and unstructured data to support AI and Agentic solutions, including pre‑processing with extraction, chunking, embedding and grounding strategies to get the data ready.
Develop AI‑driven systems to improve data capabilities, ensuring compliance with industry best practices.
Implement efficient Retrieval‑Augmented Generation (RAG) architectures and integrate them with enterprise data infrastructure.
Collaborate with cross‑functional teams to integrate solutions into operational processes and systems supporting various functions.
Stay up to date with industry advancements in AI and apply modern technologies and methodologies to our systems.
Design, build and maintain scalable and robust real‑time data streaming pipelines using technologies such as Apache Kafka, AWS Kinesis, Spark streaming or similar.
Develop data domains and data products for various consumption archetypes, including reporting, data science, AI/ML and analytics.
Ensure reliability, availability and scalability of data pipelines and systems through effective monitoring, alerting and incident management.
Implement best practices in reliability engineering, including redundancy, fault tolerance and disaster recovery strategies.
Collaborate closely with DevOps and infrastructure teams to ensure seamless deployment, operation and maintenance of data systems.
Mentor junior team members and engage in communities of practice to deliver high‑quality data and AI solutions while promoting best practices, standards and adoption of reusable patterns.
Develop graph database solutions for complex data relationships supporting AI systems.
Apply AI solutions to insurance‑specific data use cases and challenges.
Partner with architects and stakeholders to influence and implement the vision of the AI and data pipelines while safeguarding the integrity and scalability of the environment.
Qualifications
Bachelor’s or Master’s degree in Computer Science, Artificial Intelligence, or a related field.
8+ years of strong hands‑on data engineering experience, including data solutions, SQL and NoSQL, Snowflake, ETL/ELT tools, CI/CD, Big Data, cloud technologies (AWS/Google/Azure), Python/Spark, data mesh, data lake or data fabric.
Strong programming skills in Python and familiarity with deep learning frameworks such as PyTorch or TensorFlow.
Experience in implementing data governance practices, including data quality, lineage, catalogue capture, holistically, strategically and dynamically on a large‑scale data platform.
Experience with cloud platforms (AWS, GCP or Azure) and containerization technologies (Docker, Kubernetes).
Strong written and verbal communication skills and ability to explain technical concepts to various stakeholders.
Preferred Qualifications
Experience in multi‑cloud hybrid AI solutions.
AI certifications.
Experience in Employee Benefits industry.
Knowledge of natural language processing (NLP) and computer vision technologies.
Contributions to open‑source AI projects or research publications in generative AI.
Experience with building AI pipelines that bring together structured, semi‑structured and unstructured data, including pre‑processing with extraction, chunking, embedding and grounding strategies, semantic modeling and preparing the data for models and Agentic solutions.
Experience in vector databases, graph databases, NoSQL and document databases, including design, implementation and optimization (e.g., AWS OpenSearch, GCP Vertex AI, Neo4j, Spanner Graph, Neptune, Mongo, DynamoDB, etc.).
3+ years of AI/ML experience, with 1+ year of data engineering experience focused on supporting generative AI technologies.
Hands‑on experience implementing production‑ready enterprise‑grade AI data solutions.
Experience with prompt engineering techniques for large language models.
Experience in implementing Retrieval‑Augmented Generation (RAG) pipelines and integrating retrieval mechanisms with language models.
Experience in vector databases and graph databases, including implementation and optimization.
Experience in processing and leveraging unstructured data for AI applications.
Proficiency in implementing scalable AI‑driven data systems supporting agentic solution (AWS Lambda, S3, EC2, Langchain, Langgraph).
Compensation The listed annualized base pay range is: $135,040 - $202,560
Equal Opportunity Employer/Sex/Race/Color/Veterans/Disability/Sexual Orientation/Gender Identity or Expression/Religion/Age
Seniority level Mid‑Senior level
Employment type Full‑time
Job function Information Technology
About Us | Our Culture | What It’s Like to Work Here | Perks & Benefits
#J-18808-Ljbffr
Sr Staff Data Engineer - Hybrid
role at
The Hartford
Sr Staff AI Data Engineer is responsible for implementing AI data pipelines that bring together structured, semi‑structured and unstructured data to support AI and Agentic solutions. This includes pre‑processing with extraction, chunking, embedding and grounding strategies to get the data ready.
This role will have a Hybrid work schedule, with the expectation of working in an office location (Hartford, CT; Chicago, IL; Columbus, OH; and Charlotte, NC) 3 days a week (Tuesday through Thursday).
Responsibilities
AI Data Engineering lead responsible for implementing AI data pipelines that bring together structured, semi‑structured and unstructured data to support AI and Agentic solutions, including pre‑processing with extraction, chunking, embedding and grounding strategies to get the data ready.
Develop AI‑driven systems to improve data capabilities, ensuring compliance with industry best practices.
Implement efficient Retrieval‑Augmented Generation (RAG) architectures and integrate them with enterprise data infrastructure.
Collaborate with cross‑functional teams to integrate solutions into operational processes and systems supporting various functions.
Stay up to date with industry advancements in AI and apply modern technologies and methodologies to our systems.
Design, build and maintain scalable and robust real‑time data streaming pipelines using technologies such as Apache Kafka, AWS Kinesis, Spark streaming or similar.
Develop data domains and data products for various consumption archetypes, including reporting, data science, AI/ML and analytics.
Ensure reliability, availability and scalability of data pipelines and systems through effective monitoring, alerting and incident management.
Implement best practices in reliability engineering, including redundancy, fault tolerance and disaster recovery strategies.
Collaborate closely with DevOps and infrastructure teams to ensure seamless deployment, operation and maintenance of data systems.
Mentor junior team members and engage in communities of practice to deliver high‑quality data and AI solutions while promoting best practices, standards and adoption of reusable patterns.
Develop graph database solutions for complex data relationships supporting AI systems.
Apply AI solutions to insurance‑specific data use cases and challenges.
Partner with architects and stakeholders to influence and implement the vision of the AI and data pipelines while safeguarding the integrity and scalability of the environment.
Qualifications
Bachelor’s or Master’s degree in Computer Science, Artificial Intelligence, or a related field.
8+ years of strong hands‑on data engineering experience, including data solutions, SQL and NoSQL, Snowflake, ETL/ELT tools, CI/CD, Big Data, cloud technologies (AWS/Google/Azure), Python/Spark, data mesh, data lake or data fabric.
Strong programming skills in Python and familiarity with deep learning frameworks such as PyTorch or TensorFlow.
Experience in implementing data governance practices, including data quality, lineage, catalogue capture, holistically, strategically and dynamically on a large‑scale data platform.
Experience with cloud platforms (AWS, GCP or Azure) and containerization technologies (Docker, Kubernetes).
Strong written and verbal communication skills and ability to explain technical concepts to various stakeholders.
Preferred Qualifications
Experience in multi‑cloud hybrid AI solutions.
AI certifications.
Experience in Employee Benefits industry.
Knowledge of natural language processing (NLP) and computer vision technologies.
Contributions to open‑source AI projects or research publications in generative AI.
Experience with building AI pipelines that bring together structured, semi‑structured and unstructured data, including pre‑processing with extraction, chunking, embedding and grounding strategies, semantic modeling and preparing the data for models and Agentic solutions.
Experience in vector databases, graph databases, NoSQL and document databases, including design, implementation and optimization (e.g., AWS OpenSearch, GCP Vertex AI, Neo4j, Spanner Graph, Neptune, Mongo, DynamoDB, etc.).
3+ years of AI/ML experience, with 1+ year of data engineering experience focused on supporting generative AI technologies.
Hands‑on experience implementing production‑ready enterprise‑grade AI data solutions.
Experience with prompt engineering techniques for large language models.
Experience in implementing Retrieval‑Augmented Generation (RAG) pipelines and integrating retrieval mechanisms with language models.
Experience in vector databases and graph databases, including implementation and optimization.
Experience in processing and leveraging unstructured data for AI applications.
Proficiency in implementing scalable AI‑driven data systems supporting agentic solution (AWS Lambda, S3, EC2, Langchain, Langgraph).
Compensation The listed annualized base pay range is: $135,040 - $202,560
Equal Opportunity Employer/Sex/Race/Color/Veterans/Disability/Sexual Orientation/Gender Identity or Expression/Religion/Age
Seniority level Mid‑Senior level
Employment type Full‑time
Job function Information Technology
About Us | Our Culture | What It’s Like to Work Here | Perks & Benefits
#J-18808-Ljbffr