Logo
HighLevel

Lead Data Engineer

HighLevel, Dallas, Texas, United States, 75215

Save Job

Responsibilities: Software Engineering Excellence: Write clean, efficient, and maintainable code using JavaScript or Python while adhering to best practices and design patterns. Design, Build, and Maintain Systems: Develop robust software solutions and implement RESTful APIs that handle high volumes of data in real-time, leveraging message queues (Google Cloud Pub/Sub, Kafka, RabbitMQ) and event-driven architectures. Data Pipeline Development: Design, develop, and maintain data pipelines (ETL/ELT) to process structured and unstructured data from various sources. Data Storage and Warehousing: Build and optimize databases, data lakes, and data warehouses (e. g. Snowflake) for high-performance querying. Data Integration: Work with APIs, batch, and streaming data sources to ingest and transform data. Performance Optimization: Optimize queries, indexing, and partitioning for efficient data retrieval. Collaboration: Work with data analysts, data scientists, software developers, and product teams to understand requirements and deliver scalable solutions. Monitoring and Debugging: Set up logging, monitoring, and alerting to ensure data pipelines run reliably. Requirements: 4+ years of experience in software development, Bachelor's or Master's degree in Computer Science, Engineering, or a related field. Strong Problem-Solving Skills: Ability to debug and optimize data processing workflows. Programming Fundamentals: Solid understanding of data structures, algorithms, and software design patterns. Software Engineering Experience: Demonstrated experience (SDE II/III level) in designing, developing, and delivering software solutions using modern languages and frameworks (Node.js, JavaScript, Python, TypeScript, SQL, Scala, or Java). ETL Tools and Frameworks: Experience with Airflow, dbt, Apache Spark, Kafka, Flink, or similar technologies. Cloud Platforms: Hands-on experience with GCP (Pub/Sub, Dataflow, Cloud Storage) or AWS (S3 Glue, Redshift). Databases and Warehousing: Strong experience with PostgreSQL, MySQL, Snowflake, and NoSQL databases (MongoDB, Firestore, ES). Version Control and CI/CD: Familiarity with Git, Jenkins, Docker, Kubernetes, and CI/CD pipelines for deployment. Communication: Excellent verbal and written communication skills, with the ability to work effectively in a collaborative environment. Experience with data visualization tools (e. g, Superset, Tableau), Terraform, IaC, ML/AI data pipelines, and devops practices is a plus.

#J-18808-Ljbffr