We are seeking a highly skilled and motivated Advanced Data Analytics Engineer to join our data team. In this role, you will be responsible for developing, implementing, and optimizing advanced data analytics solutions to drive business insights and decision-making. You will work closely with data scientists, data engineers, and business analysts to build and maintain data pipelines, machine learning models, and analytics tools that provide actionable insights from complex datasets. This is an excellent opportunity for an engineer passionate about working with large-scale data and advanced analytics to solve real-world business challenges.
Roles and Responsibilities:
Design and develop scalable data pipelines for advanced analytics and reporting.
Build and optimize data models for clustering, segmentation, and scoring tasks.
Integrate data from multiple sources using Kafka , Airflow , and ETL tools.
Support data governance and quality checks to ensure data integrity and accuracy.
Collaborate closely with the Data Analytics Lead and AI/ML teams for the integration of data into advanced analytics and AI-driven products.
Manage and optimize storage solutions using MinIO or S3 , ensuring data is properly partitioned and indexed for performance.
Must Have Skills:
7+ years of experience in data engineering and analytics.
Strong experience with cloud data solutions (AWS preferred), ETL processes , and Big Data tools .
Hands-on experience with data storage systems such as S3 , MinIO , and Snowflake .
Familiarity with Kafka for real-time data streaming and Airflow for scheduling and orchestration.
Proficiency in data pipelines (e.g., Airflow , Kafka , ETL frameworks ).
Strong experience in data storage systems like Snowflake , Hive , Iceberg , S3 , and MinIO .
Solid understanding of data modeling , segmentation , and clustering techniques.
Advanced knowledge of SQL and relational/non-relational databases.
Strong programming skills in Python or Java for data engineering.
Good to Have:
Familiarity with AWS services such as S3, Redshift, Lambda, and EMR.
Experience with DataOps , and using DevOps tools for automation in data pipelines.
Qualification:
Bachelor of Science in Computer Science or equivalent technical training and professional work experience