AuxoAI is seeking a skilled and experienced Data Engineer to join our dynamic team. The ideal candidate will have 3-6 years of prior experience in data engineering, with a strong background in AWS (Amazon Web Services) technologies. This role offers an exciting opportunity to work on diverse projects, collaborating with cross-functional teams to design, build, and optimize data pipelines and infrastructure.
Additionally, you are required to stay up to date on AI trends and best practices, sharing knowledge with the team to foster continuous learning and improvement.
If you are proactive, self-motivated, and passionate about AI, this role offers an exciting opportunity to make a meaningful impact and drive business transformation.
Responsibilities
Design, develop, and maintain scalable data pipelines and ETL processes leveraging AWS services such as S3, Glue, EMR, Lambda, Aurora, RDS, Lake formation, Athena, DMS and Redshift
Collaborate with data scientists and analysts to understand data requirements and implement solutions that support analytics and machine learning initiatives
Optimize data storage and retrieval mechanisms to ensure performance, reliability, and cost-effectiveness
Implement data governance and security best practices to ensure compliance and data integrity
Troubleshoot and debug data pipeline issues, providing timely resolution and proactive monitoring
Stay abreast of emerging technologies and industry trends, recommending innovative solutions to enhance data engineering capabilities
Develop and maintain strong relationships with key clients, serving as a trusted advisor and strategic partner while identifying opportunities for upselling and cross-selling additional services to drive revenue growth
Qualifications
Bachelors or Masters degree in Computer Science, Engineering, or a related field
3-6 years of prior experience in data engineering, with a focus on designing and building data pipelines
Proficiency in AWS services, particularly S3, Glue, EMR, Lambda, Aurora, RDS, MWAA, Lake formation, Athena, DMS and Redshift
Strong programming skills in languages such as Python, Java, or Scala
Proficient in Spark, Databricks and messaging queues like RabbitMQ and Kafka
Experience with SQL and NoSQL databases, data warehousing concepts, and big data technologies
Familiarity with containerization technologies (e.g., Docker, Kubernetes) and orchestration tools (e.g., Apache Airflow)