Logo
Intelliswift Software Inc

AWS Data Engineer

Intelliswift Software Inc, Torrance, California, United States, 90504

Save Job

Job ID: 25-10803 Daily Tasks Performed: Develop and Maintain Data Integration Solutions:

Design

and

implement data integration workflows

using

AWS Glue EMR

,

Lambda

,

Redshift Pyspark

,

Spark

and

Python

for

data processing large datasets Ensure data is

extracted, transformed, and loaded into target systems Build ETL pipelines using Iceberg Ensure Data Quality and Integrity:

Validate

and

cleanse data Ensure data quality

and

integrity

by

implementing monitoring

,

validation

, and

error handling mechanisms

within

data pipelines Optimize Data Integration Processes:

Enhance the

performance

,

optimization

of

data workflows

to

meet SLAs

, scalability of data integration

on AWS cloud infrastructure Data Analysis

and

Data Warehousing concepts

(

star snowflake schema design

,

dimensional modeling

, and

reporting enablement

) Resolve performance bottlenecks Optimize data processing to enhance Redshift's performance Refine integration processes Support Business Intelligence and Analytics:

Translate business requirements

to

technical specifications

and

coded data pipelines Ensure

integrated data for business intelligence

and

analytics Meet data requirements Maintain Documentation and Compliance:

Document all data integration processes

,

workflows

, and technical & system specifications. Ensure

compliance with data governance policies

, industry standards, and regulatory requirements. What will this person be working on

Design

,

development

, and

management

of

data integration processes Integrating data from diverse sources

,

transforming it to meet business requirements

, and

loading it into target systems

such as

data warehouses

or

data lakes Position Success Criteria (Desired) • 'WANTS'

Bachelor's degree

in computer science, information technology, or a related field. A master's degree can be advantageous. 4-6+ years of experience

in

data engineering

,

database design

,

ETL processes Experience with

Iceberg 5+

in

programming languages

such as

PySpark

,

Python and SQL 5+ years of experience

with

AWS tools and technologies

(

S3

,

EMR

,

Glue

,

Athena

,

RedShift

,

Postgres

,

RDS

,

Lambda

,

PySpark

) 3+ years of experience of working with databases data marts data warehouses ETL development

,

system integration

, and

CI CD implementation Experience in

complex database objects

to move the changed data across multiple environments Solid understanding of

data security

, privacy, and compliance. Participate in

agile development processes

including sprint planning stand-ups and retrospectives Provide technical guidance and mentorship to junior developers