Leidos Inc
Overview
Leidos is seeking a Databricks Data Engineer to develop/support new/existing data pipelines and data analytics environments in an Azure cloud-based data lake. As a data engineer, you will translate business requirements to data engineering solutions to support an enterprise-scale Microsoft Azure based data analytics platform. You will support continued maintenance of ETL operations and development of new pipelines ensuring data quality and data management. The ideal candidate will bring deep expertise in Databricks, a solid foundation in advanced AI technologies, and apply critical thinking to create innovative functions and solve technical issues by cross-functional team collaboration. Note:
This role is remote for candidates located within the United States; however, the ideal candidate will be based in or willing to relocate to the Washington, DC or Indianapolis, IN area to support periodic on-site activities. Responsibilities
Design, build, and optimize scalable data solutions using Databricks and Medallion Architecture. Manage ingestion routines for processing multi-terabyte datasets efficiently for multiple projects simultaneously, where each project may have multiple Databricks workspaces. Integrate data from various structured and unstructured sources to enable high-quality business insights. Proficiency in data analysis techniques for deriving insights from large datasets. Implement effective data management strategies to ensure data integrity, availability, and accessibility. Identify opportunities for cost optimization in data storage, processing, and analytics operations. Monitor and support user requests, addressing platform or performance issues, cluster stability, Spark optimization, and configuration management. Collaborate with the team to enable advanced AI-driven analytics and data science workflows. Integrate with Azure services including Azure Functions, Storage Services, Data Factory, Log Analytics, and User Management for seamless data workflows. Experience with these Azure services is a plus. Provision and manage infrastructure using Infrastructure-as-Code (IaC). Apply best practices for data security, data governance, and compliance, ensuring support for federal regulations and public trust standards. Proactively collaborate with technical and non-technical teams to gather requirements and translate business needs into data solutions. Qualifications
BS degree in Computer Science or related field and 3+ years of experience or Master's degree with 2+ years of experience. 3+ years of experience developing and designing ingestion flows (structured, streaming, and unstructured data) using cloud platform services with data quality. Databricks Data Engineer certification and 2+ years of experience maintaining Databricks platform and development in Spark. Ability to work directly with clients and act as front-line support for requests coming in from clients. Clearly document and express the solution in the form of architecture and interface diagrams. Proficient in Python, Spark, and R. .NET based development is a plus. Knowledge and experience with data governance, including metadata management, enterprise data catalog, design standards, data quality governance, and data security. Experience with Agile methodology, CI/CD automation, and cloud-based developments (Azure, AWS). Not required, but additional education, certifications, and/or experience are a plus: Azure cloud certifications, knowledge of FinOps principles and cost management. Pay Range:
$85,150.00 - $153,925.00
#J-18808-Ljbffr
Leidos is seeking a Databricks Data Engineer to develop/support new/existing data pipelines and data analytics environments in an Azure cloud-based data lake. As a data engineer, you will translate business requirements to data engineering solutions to support an enterprise-scale Microsoft Azure based data analytics platform. You will support continued maintenance of ETL operations and development of new pipelines ensuring data quality and data management. The ideal candidate will bring deep expertise in Databricks, a solid foundation in advanced AI technologies, and apply critical thinking to create innovative functions and solve technical issues by cross-functional team collaboration. Note:
This role is remote for candidates located within the United States; however, the ideal candidate will be based in or willing to relocate to the Washington, DC or Indianapolis, IN area to support periodic on-site activities. Responsibilities
Design, build, and optimize scalable data solutions using Databricks and Medallion Architecture. Manage ingestion routines for processing multi-terabyte datasets efficiently for multiple projects simultaneously, where each project may have multiple Databricks workspaces. Integrate data from various structured and unstructured sources to enable high-quality business insights. Proficiency in data analysis techniques for deriving insights from large datasets. Implement effective data management strategies to ensure data integrity, availability, and accessibility. Identify opportunities for cost optimization in data storage, processing, and analytics operations. Monitor and support user requests, addressing platform or performance issues, cluster stability, Spark optimization, and configuration management. Collaborate with the team to enable advanced AI-driven analytics and data science workflows. Integrate with Azure services including Azure Functions, Storage Services, Data Factory, Log Analytics, and User Management for seamless data workflows. Experience with these Azure services is a plus. Provision and manage infrastructure using Infrastructure-as-Code (IaC). Apply best practices for data security, data governance, and compliance, ensuring support for federal regulations and public trust standards. Proactively collaborate with technical and non-technical teams to gather requirements and translate business needs into data solutions. Qualifications
BS degree in Computer Science or related field and 3+ years of experience or Master's degree with 2+ years of experience. 3+ years of experience developing and designing ingestion flows (structured, streaming, and unstructured data) using cloud platform services with data quality. Databricks Data Engineer certification and 2+ years of experience maintaining Databricks platform and development in Spark. Ability to work directly with clients and act as front-line support for requests coming in from clients. Clearly document and express the solution in the form of architecture and interface diagrams. Proficient in Python, Spark, and R. .NET based development is a plus. Knowledge and experience with data governance, including metadata management, enterprise data catalog, design standards, data quality governance, and data security. Experience with Agile methodology, CI/CD automation, and cloud-based developments (Azure, AWS). Not required, but additional education, certifications, and/or experience are a plus: Azure cloud certifications, knowledge of FinOps principles and cost management. Pay Range:
$85,150.00 - $153,925.00
#J-18808-Ljbffr