We are looking for a Data Engineer with experience in Python , Spark SQL , Hive , and Airflow . In this role, you will be responsible for building and maintaining data pipelines, processing large datasets, and automating workflows to streamline data-related tasks.
Required Skills Experience :
Mandatory :
Strong programming experience in Python for data engineering tasks.
Expertise in Spark SQL for big data querying and processing.
Hands-on experience with Apache Hive for data warehousing and querying.
Proficiency in building, scheduling, and managing data pipelines using Apache Airflow .
Experience working with large datasets in distributed environments like Hadoop or Spark clusters.
Strong knowledge of SQL for data manipulation and extraction.
Familiarity with cloud environments like AWS , Azure , or GCP .
Knowledge of version control systems (e.g., Git ).
Good to Have :
Experience with other big data processing tools such as Apache Kafka , Flink , or Presto .
Knowledge of data modeling and designing data lakes and data warehouses.
Experience in DataOps and CI/CD processes for data pipelines.
Familiarity with Docker or Kubernetes for containerization and orchestration of data processing tasks.
Experience in performance tuning of big data systems and optimizing complex queries.