Upload Button Icon Add office photos
Engaged Employer

i

This company page is being actively managed by Cognizant Team. If you also belong to the team, you can get access from here

Cognizant Verified Tick

Compare button icon Compare button icon Compare
3.8

based on 47.8k Reviews

Proud winner of ABECA 2024 - AmbitionBox Employee Choice Awards

zig zag pattern zig zag pattern
filter salaries All Filters

4 Cognizant Pyspark Developer Jobs

Pyspark Developer- Indore

4-12 years

Bangalore / Bengaluru

1 vacancy

Pyspark Developer- Indore

Cognizant

posted 7d ago

Job Role Insights

Flexible timing

Job Description

Job Title PySpark Developer

Location Indore

Job Type Full time

Years of Experience 4 to 12 years.

Job Description

We are seeking an experienced PySpark Developer to join our data engineering team. In this role, you will be responsible for designing, developing, and optimizing large-scale data processing pipelines using PySpark and other big data technologies. The ideal candidate will have expertise in distributed computing, data processing frameworks, and working with large datasets in cloud-based or on-premises environments. You will collaborate with data engineers, data scientists, and business analysts to build robust, scalable, and efficient data solutions.

Key Responsibilities

Data Processing & Transformation Design, develop, and implement distributed data processing and transformation workflows using PySpark to handle large-scale datasets across various storage systems (HDFS, S3, etc. ).

ETL Development Build and manage ETL (Extract, Transform, Load) pipelines using PySpark, integrating data from multiple sources such as databases, flat files, cloud storage, and other data platforms.

Data Wrangling & Cleansing Perform data cleaning, data wrangling, and data transformations to ensure the integrity, accuracy, and completeness of the data before feeding it into analytical models or reports.

Optimization & Performance Tuning Optimize PySpark jobs for better performance, such as minimizing memory usage, optimizing partitioning, and tuning Spark configurations for faster data processing.

Collaboration with Data Scientists Work closely with data scientists to help preprocess large datasets, manage data pipelines, and support machine learning model deployment and experimentation.

Big Data Technologies Integration Integrate PySpark with other big data technologies (e. g. , Hadoop, Hive, Kafka, NoSQL databases) to process structured and unstructured data in real-time or batch modes.

Data Modeling Work with data engineers to design and implement data models that support efficient storage and querying, ensuring data can be leveraged for analytics, BI, and machine learning use cases.

Testing & Debugging Ensure the accuracy and reliability of data processing by conducting unit tests, integration tests, and debugging PySpark jobs in a distributed environment.

Documentation Create and maintain documentation for PySpark applications, data workflows, and procedures to ensure clarity and knowledge transfer across teams.

Monitoring & Support Monitor data pipelines and jobs, ensuring they run efficiently and handle exceptions or errors effectively. Provide support for production systems as needed.

Required Skills and Qualifications

PySpark Expertise Strong experience with PySpark for developing distributed data processing workflows, transformations, and optimizations on large datasets.

Big Data Frameworks Proficiency with big data technologies such as Hadoop, Hive, Spark, Kafka, or other distributed processing frameworks.

Programming Skills Solid knowledge of Python for data manipulation, scripting, and automating tasks. Familiarity with other languages like Scala or Java is a plus.

SQL Skills Proficient in SQL for querying databases and integrating with PySpark to extract and manipulate structured data.

Data Storage Experience with cloud storage systems (e. g. , Amazon S3, Azure Blob Storage) and distributed file systems (e. g. , HDFS).

Data Processing & Integration Experience in building data pipelines and integrating disparate data sources for processing, analysis, and reporting.

Performance Tuning & Troubleshooting Expertise in optimizing PySpark jobs for performance and troubleshooting issues in a distributed computing environment.

Cloud Platforms Experience working with cloud platforms like AWS, Azure, or Google Cloud, specifically their big data offerings (e. g. , AWS EMR, Azure Databricks, Google Dataproc).

Version Control Familiarity with Git or other version control tools for collaborative development and deployment.

Problem-Solving Strong analytical skills with the ability to break down complex problems

We are seeking a highly skilled Sr. Developer with 4 to 8 years of experience to join our team. The ideal candidate will have expertise in Python Databricks SQL Databricks Workflows and PySpark. Experience in Park Operations is a plus. This role involves developing and optimizing data workflows to support our business objectives and enhance operational efficiency.

Responsibilities

  • Develop and maintain data workflows using Databricks Workflows to ensure seamless data integration and processing.
  • Utilize Python to create efficient and scalable data processing scripts.
  • Implement and optimize SQL queries within Databricks to support data analysis and reporting needs.
  • Leverage PySpark to handle large-scale data processing tasks and improve performance.
  • Collaborate with cross-functional teams to understand business requirements and translate them into technical solutions.
  • Provide technical guidance and support to junior developers to foster skill development and knowledge sharing.
  • Conduct code reviews to ensure code quality and adherence to best practices.
  • Troubleshoot and resolve technical issues related to data workflows and processing.
  • Monitor and optimize the performance of data workflows to ensure they meet business requirements.
  • Develop and maintain documentation for data workflows processes and best practices.
  • Stay updated with the latest industry trends and technologies to continuously improve data processing capabilities.
  • Work closely with stakeholders to gather requirements and provide regular updates on project progress.
  • Ensure data security and compliance with company policies and industry regulations.

Qualifications

  • Possess strong experience in Python for data processing and automation.
  • Demonstrate expertise in Databricks SQL for data analysis and reporting.
  • Have hands-on experience with Databricks Workflows for data integration.
  • Show proficiency in PySpark for large-scale data processing.
  • Experience in Park Operations is a plus providing valuable domain knowledge.
  • Exhibit excellent problem-solving skills and attention to detail.
  • Display strong communication skills to effectively collaborate with team members and stakeholders.
  • Have a proactive approach to learning and staying updated with new technologies.

Certifications Required

Databricks Certified Associate Developer for Apache Spark Python Certification

The Cognizant community
We are a high caliber team who appreciate and support one another. Our people uphold an energetic, collaborative and inclusive workplace where everyone can thrive.

  • Cognizant is a global community with more than 350, 000 associates around the world.
  • We don t just dream of a better way - we make it happen.
  • We take care of our people, clients, company, communities and climate by doing what s right.
  • We foster an innovative environment where you can build the career path that s right for you.

About us
Cognizant is one of the world's leading professional services companies, transforming clients' business, operating, and technology models for the digital era. Our unique industry-based, consultative approach helps clients envision, build, and run more innovative and efficient businesses. Headquartered in the U. S. , Cognizant (a member of the NASDAQ-100 and one of Forbes World s Best Employers 2024) is consistently listed among the most admired companies in the world. Learn how Cognizant helps clients lead with digital at www. cognizant. com

Our commitment to diversity and inclusion

CareersNA2@cognizant. com with your request and contact information.

Disclaimer


Employment Type: Full Time, Permanent

Read full job description

Cognizant Interview Questions & Tips

Prepare for Cognizant Pyspark Developer roles with real interview advice

Top Cognizant Pyspark Developer Interview Questions

Q1. What is the difference between coalesce and repartition, as well as between cache and persist?
Q2. What is the SQL query to find the second highest rank in a dataset?
Q3. What is the SQL code for calculating year-on-year growth percentage with year-wise grouping?
View all 8 questions

What Pyspark Developer at Cognizant are saying

4.0
 Rating based on 1 Pyspark Developer review

Likes

Good Work Environment and culture. Team mates are supportive.

Read 1 review

Pyspark Developer salary at Cognizant

reported by 22 employees with 1-4 years exp.
₹4 L/yr - ₹9 L/yr
16% less than the average Pyspark Developer Salary in India
View more details

What Cognizant employees are saying about work life

based on 47.8k employees
66%
87%
70%
73%
Flexible timing
Monday to Friday
No travel
Day Shift
View more insights

Cognizant Benefits

Submitted by Company
Shape your career
Invest in yourself
Learn and grow
Keep current
Stay healthy
Give back
Submitted by Employees
Health Insurance
Cafeteria
Work From Home
Free Transport
Soft Skill Training
Job Training +6 more
View more benefits

Compare Cognizant with

TCS

3.7
Compare

Infosys

3.7
Compare

Wipro

3.7
Compare

Accenture

3.9
Compare

Capgemini

3.8
Compare

Tech Mahindra

3.6
Compare

IBM

4.1
Compare

Deloitte

3.8
Compare

Amazon

4.1
Compare

Google

4.4
Compare

PwC

3.4
Compare

Persistent Systems

3.5
Compare

Ernst & Young

3.5
Compare

DXC Technology

3.7
Compare

HCLTech

3.5
Compare

Virtusa Consulting Services

3.8
Compare

KPMG India

3.5
Compare

Zoho

4.3
Compare

Nagarro

4.0
Compare

Oracle

3.7
Compare

Similar Jobs for you

Snowflake Developer at Cognizant Technology Solutions India Ltd

Chennai

4-12 Yrs

₹ 8-9 LPA

Pyspark Developer at Newt Global

Mumbai

6-8 Yrs

₹ 8-10 LPA

Pyspark Developer at Infosys

Hyderabad / Secunderabad, Pune + 1

6-11 Yrs

₹ 7-17 LPA

Pyspark Developer at Infosys Limited

Bangalore / Bengaluru

5-7 Yrs

₹ 7-9 LPA

Pyspark Developer at Infosys Limited

Bangalore / Bengaluru

5-7 Yrs

₹ 7-9 LPA

Pyspark Developer at DATAECONOMY

Chennai, Pune + 5

2-6 Yrs

₹ 4-8 LPA

Pyspark Developer at Fusion Plus Solutions Inc

Hyderabad / Secunderabad

2-5 Yrs

₹ 4-7 LPA

Pyspark Developer at Infosys Limited

Bangalore / Bengaluru

3-5 Yrs

₹ 5-7 LPA

Pyspark Developer at Overture Rede pvt ltd

Bangalore / Bengaluru

3-10 Yrs

₹ 3-7 LPA

Pyspark Developer at Capgemini Technology Services India Limited

Mumbai, Pune + 1

6-11 Yrs

₹ 8-14 LPA

Cognizant Bangalore / Bengaluru Office Location

View all
Bengaluru Office
Cognizant, Baghmane Tech Park, 65/2 -1, Adjacent LRDE, Byrasandra, C.V.Raman Nagar Bengaluru
Karnataka 560093

Pyspark Developer- Indore

4-12 Yrs

Bangalore / Bengaluru

7d ago·via naukri.com

Walkin || Cognizant is hiring For Pyspark Developers

6-11 Yrs

Chennai

8d ago·via naukri.com

Pyspark Developer

7-9 Yrs

Pan inida

8d ago·via naukri.com

Pyspark Developer

4-8 Yrs

Hyderabad / Secunderabad, Pune, Bangalore Rural

26d ago·via naukri.com

Similar Designation Jobs in Cognizant

Pyspark Developer Jobs in Other Companies

write
Share an Interview