Create and implement automated ETL tasks from business applications and public data sources (such as census data) into a modern data lakehouse
Implement a cutting-edge data lakehouse solution while advising the lead architect on technology and architecture based on use cases and intended results
As part of a systems migration for business operations, support data migration from an aging SQL server to a new RDBMS
Ensure that the data lakehouse contains the components for the data catalog and associated documentation for analytical end users
Transparently discuss deadlines and deliverables with the team and leadership
Fulfill deadlines and collaborate with IT leadership on iterative development
Job Requirements:
Bachelor s/Master s degree in Engineering, Computer Science (or equivalent experience)
At least 4+ years of relevant experience with application development or as a data engineer
4+ years of experience working with Python, SQL, and Scala
3+ years of experience with AWS and related technologies
3+ years of experience using distributed computing and data tools including MapReduce, Hadoop, Hive, EMR, Spark, and Athena
3+ years of working experience with ETL tools like Kafka, AirFlow, and SSIS
In-depth experience developing data lake or data warehouse solutions like Snowflake and Redshift
Prolific experience working with Git version control and GitLab/GitHub remote repositories
Ability to work autonomously, troubleshoot, and communicate proactively
Nice to have some experience working in agile and iterative development environments
Prior experience interfacing with .NET API (.NET C# for .NET Core 6) is desirable
Some familiarity with Elasticsearch and/or Solr is preferred
Excellent verbal and written English communication skills