AHU Technologies Inc
16 years Big Data Architect IT Consultant with Databricks
AHU Technologies Inc, Washington, District of Columbia, us, 20022
Overview
Role: Big Data Architect IT Consultant Master. Client: State of DC. Location: Washington, D.C.
This role will provide expertise to support the development of a Big Data / Data Lake system architecture that supports enterprise data operations for the District of Columbia government, including IoT / Smart City projects, enterprise data warehouse, open data portal, and data science applications. The architecture includes Databricks, Microsoft Azure platform tools (including Data Lake, Synapse), Apache tools (including Hadoop, Hive, Impala, Spark, Sedona, Airflow) and data pipeline/ETL tools (including Streamsets, Apache NiFi, Azure Data Factory). The platform will be designed for District-wide use and integration with other OCTO Enterprise Data tools such as Esri, Tableau, MicroStrategy, API Gateways, and Oracle databases and integration tools.
Responsibilities
Coordinates IT project management, engineering, maintenance, QA, and risk management.
Plans, coordinates, and monitors project activities.
Develops technical applications to support users.
Develops, implements, maintains and enforces documented standards and procedures for the design, development, installation, modification, and documentation of assigned systems.
Provides training for system products and procedures.
Performs application upgrades.
Performs monitoring, maintenance, or reporting on real-time databases, real-time network and serial data communications, and real-time graphics and logic applications.
Troubleshoots problems.
Ensures project life-cycle is in compliance with District standards and procedures.
Minimum Education/Certification Requirements Bachelor’s degree in Information Technology or related field or equivalent experience
Skills
Experience implementing Big Data storage and analytics platforms such as Databricks and Data Lakes
Knowledge of Big Data and Data Architecture and Implementation best practices — 5 Years
Knowledge of architecture and implementation of networking, security and storage on cloud platforms such as Microsoft Azure — 5 Years
Experience with deployment of data tools and storage on cloud platforms such as Microsoft Azure — 5 Years
Knowledge of Data-centric systems for the analysis and visualization of data, such as Tableau, MicroStrategy, ArcGIS, Kibana, Oracle — 10 Years
Experience querying structured and unstructured data sources including SQL and NoSQL databases — 5 Years
Experience modeling and ingesting data into and between various data systems through the use of Data Pipelines — 5 Years
Experience with implementing Apache data products such as Spark, Sedona, Airflow, Atlas, NiFi, Hive, Impala — 5 Years
Experience with API / Web Services (REST/SOAP) — 3 Years
Experience with complex event processing and real-time streaming data — 3 Years
Experience with deployment and management of data science tools and modules such as JupyterHub — 3 Years
Experience with ETL, data processing, analytics using languages such as Python, Java or R — 3 Years
Experience with Cloudera Data Platform — 3 Years
16+ years planning, coordinating, and monitoring project activities
16+ years leading projects, ensuring they are in compliance with established standards/procedures
Bachelor’s degree in IT or related field or equivalent experience
Required
#J-18808-Ljbffr
This role will provide expertise to support the development of a Big Data / Data Lake system architecture that supports enterprise data operations for the District of Columbia government, including IoT / Smart City projects, enterprise data warehouse, open data portal, and data science applications. The architecture includes Databricks, Microsoft Azure platform tools (including Data Lake, Synapse), Apache tools (including Hadoop, Hive, Impala, Spark, Sedona, Airflow) and data pipeline/ETL tools (including Streamsets, Apache NiFi, Azure Data Factory). The platform will be designed for District-wide use and integration with other OCTO Enterprise Data tools such as Esri, Tableau, MicroStrategy, API Gateways, and Oracle databases and integration tools.
Responsibilities
Coordinates IT project management, engineering, maintenance, QA, and risk management.
Plans, coordinates, and monitors project activities.
Develops technical applications to support users.
Develops, implements, maintains and enforces documented standards and procedures for the design, development, installation, modification, and documentation of assigned systems.
Provides training for system products and procedures.
Performs application upgrades.
Performs monitoring, maintenance, or reporting on real-time databases, real-time network and serial data communications, and real-time graphics and logic applications.
Troubleshoots problems.
Ensures project life-cycle is in compliance with District standards and procedures.
Minimum Education/Certification Requirements Bachelor’s degree in Information Technology or related field or equivalent experience
Skills
Experience implementing Big Data storage and analytics platforms such as Databricks and Data Lakes
Knowledge of Big Data and Data Architecture and Implementation best practices — 5 Years
Knowledge of architecture and implementation of networking, security and storage on cloud platforms such as Microsoft Azure — 5 Years
Experience with deployment of data tools and storage on cloud platforms such as Microsoft Azure — 5 Years
Knowledge of Data-centric systems for the analysis and visualization of data, such as Tableau, MicroStrategy, ArcGIS, Kibana, Oracle — 10 Years
Experience querying structured and unstructured data sources including SQL and NoSQL databases — 5 Years
Experience modeling and ingesting data into and between various data systems through the use of Data Pipelines — 5 Years
Experience with implementing Apache data products such as Spark, Sedona, Airflow, Atlas, NiFi, Hive, Impala — 5 Years
Experience with API / Web Services (REST/SOAP) — 3 Years
Experience with complex event processing and real-time streaming data — 3 Years
Experience with deployment and management of data science tools and modules such as JupyterHub — 3 Years
Experience with ETL, data processing, analytics using languages such as Python, Java or R — 3 Years
Experience with Cloudera Data Platform — 3 Years
16+ years planning, coordinating, and monitoring project activities
16+ years leading projects, ensuring they are in compliance with established standards/procedures
Bachelor’s degree in IT or related field or equivalent experience
Required
#J-18808-Ljbffr