41 Hanker Systems Jobs
PySpark Developer
Hanker Systems
posted 2d ago
Key skills for the job
We are seeking an experienced PySpark Developer / Data Engineer to design, develop, and optimize big data processing pipelines using Apache Spark and Python (PySpark). The ideal candidate should have expertise in distributed computing, ETL workflows, data lake architectures, and cloud-based big data solutions.
Key Responsibilities:
Develop and optimize ETL/ELT data pipelines using PySpark on distributed computing platforms (Hadoop, Databricks, EMR, HDInsight).
Work with structured and unstructured data to perform data transformation, cleansing, and aggregation.
Implement data lake and data warehouse solutions on AWS (S3, Glue, Redshift), Azure (ADLS, Synapse), or GCP (BigQuery, Dataflow).
Optimize PySpark jobs for performance tuning, partitioning, and caching strategies.
Design and implement real-time and batch data processing solutions.
Integrate data pipelines with Kafka, Delta Lake, Iceberg, or Hudi for streaming and incremental updates.
Ensure data security, governance, and compliance with industry best practices.
Work with data scientists and analysts to prepare and process large-scale datasets for machine learning models.
Collaborate with DevOps teams to deploy, monitor, and scale PySpark jobs using CI/CD pipelines, Kubernetes, and containerization.
Perform unit testing and validation to ensure data integrity and reliability.
Required Skills & Qualifications:
6+ years of experience in big data processing, ETL, and data engineering.
Strong hands-on experience with PySpark (Apache Spark with Python).
Expertise in SQL, DataFrame API, and RDD transformations.
Experience with big data platforms (Hadoop, Hive, HDFS, Spark SQL).
Knowledge of cloud data processing services (AWS Glue, EMR, Databricks, Azure Synapse, GCP Dataflow).
Proficiency in writing optimized queries, partitioning, and indexing for performance tuning.
Experience with workflow orchestration tools like Airflow, Oozie, or Prefect.
Familiarity with containerization and deployment using Docker, Kubernetes, and CI/CD pipelines.
Strong understanding of data governance, security, and compliance (GDPR, HIPAA, CCPA, etc.).
Excellent problem-solving, debugging, and performance optimization skills.
Employment Type: Full Time, Permanent
Read full job descriptionPrepare for Pyspark Developer roles with real interview advice