Capgemini
Technical Data Architect at New York, NY / Tampa, FL - Onsite - Fulltime
This range is provided by Capgemini. Your actual pay will be based on your skills and experience — talk with your recruiter to learn more.
Base pay range $130,000.00/yr - $135,000.00/yr
Direct message the job poster from Capgemini
Responsibilities:
Architect streaming data ingestion and integration with downstream systems
Implement AI-driven controller to orchestrate tens of millions of streams and micro-batches
Design AI-powered onboarding of new data sources
Develop AI-powered compute engine and data serving semantic layer
Deliver scalable cloud data services and APIs with sub-second response times over petabytes of data
Develop a unified alerting and monitoring framework supporting streaming transformations and compute across thousands of institutional clients and hundreds of external data sources
Build a self-service data management and operations platform
Implement a data quality monitoring framework
Qualifications:
Bachelor’s degree in Computer Science, related field; advanced degree preferred
12+ years of experience in data technology
At least 5 years as a Data Engineer with hands‑on experience in cloud environments
8+ years of Python programming focused on data processing and distributed systems
8+ years working with relational databases, SQL, dimensional modeling, and DBT
8+ years designing and administering cloud-based data warehousing solutions (e.g., Snowflake, Databricks)
8+ years experience with Kafka or other streaming platforms
Exposure to AI based advance techniques and tools
Strong understanding of database fundamentals, including data modeling, advanced SQL development and optimization, ELT/ETL processes and DBT.
Experience with Java, Oracle, MS SQL Server, Druid, Qlik/Golden Gate CDC, and Power BI is a plus
Strong leadership abilities and excellent communication skills
#J-18808-Ljbffr
Base pay range $130,000.00/yr - $135,000.00/yr
Direct message the job poster from Capgemini
Responsibilities:
Architect streaming data ingestion and integration with downstream systems
Implement AI-driven controller to orchestrate tens of millions of streams and micro-batches
Design AI-powered onboarding of new data sources
Develop AI-powered compute engine and data serving semantic layer
Deliver scalable cloud data services and APIs with sub-second response times over petabytes of data
Develop a unified alerting and monitoring framework supporting streaming transformations and compute across thousands of institutional clients and hundreds of external data sources
Build a self-service data management and operations platform
Implement a data quality monitoring framework
Qualifications:
Bachelor’s degree in Computer Science, related field; advanced degree preferred
12+ years of experience in data technology
At least 5 years as a Data Engineer with hands‑on experience in cloud environments
8+ years of Python programming focused on data processing and distributed systems
8+ years working with relational databases, SQL, dimensional modeling, and DBT
8+ years designing and administering cloud-based data warehousing solutions (e.g., Snowflake, Databricks)
8+ years experience with Kafka or other streaming platforms
Exposure to AI based advance techniques and tools
Strong understanding of database fundamentals, including data modeling, advanced SQL development and optimization, ELT/ETL processes and DBT.
Experience with Java, Oracle, MS SQL Server, Druid, Qlik/Golden Gate CDC, and Power BI is a plus
Strong leadership abilities and excellent communication skills
#J-18808-Ljbffr