Premium Employer

i

This company page is being actively managed by Persistent Systems Team. If you also belong to the team, you can get access from here

Persistent Systems Verified Tick

Compare button icon Compare button icon Compare

Filter interviews by

Persistent Systems Data Engineer Interview Questions and Answers

Updated 26 Feb 2024

Persistent Systems Data Engineer Interview Experiences

2 interviews found

Data Engineer Interview Questions & Answers

user image Anonymous

posted on 15 Feb 2024

Interview experience
3
Average
Difficulty level
Moderate
Process Duration
2-4 weeks
Result
Selected Selected

I applied via Naukri.com

Round 1 - Technical 

(2 Questions)

  • Q1. Aws sevices you have worked on
  • Ans. 

    I have worked on various AWS services including S3, EC2, Lambda, Glue, and Redshift.

    • S3 - Used for storing and retrieving data

    • EC2 - Used for running virtual servers

    • Lambda - Used for serverless computing

    • Glue - Used for ETL (Extract, Transform, Load) processes

    • Redshift - Used for data warehousing and analytics

  • Answered by AI
  • Q2. Question on python list and dictionary
Round 2 - Technical 

(1 Question)

  • Q1. Questions on project and skillset

Interview Preparation Tips

Interview preparation tips for other job seekers - 3 rounds basically asked in your resume. Questions matching skillset and project are asked

Skills evaluated in this interview

Data Engineer Interview Questions & Answers

user image Anonymous

posted on 26 Feb 2024

Interview experience
3
Average
Difficulty level
-
Process Duration
-
Result
Not Selected

I applied via Naukri.com and was interviewed in Jan 2024. There was 1 interview round.

Round 1 - One-on-one 

(4 Questions)

  • Q1. Lazy evaluation narrow wide tranform data brick utlity functions
  • Q2. Dag,edges vertices
  • Q3. Spark architecture
  • Q4. Memory allocation

Data Engineer Interview Questions Asked at Other Companies

asked in Cisco
Q1. Optimal Strategy for a Coin Game You are playing a coin game with ... read more
asked in Sigmoid
Q2. Next Greater Element Problem Statement You are given an array arr ... read more
asked in Sigmoid
Q3. Problem: Search In Rotated Sorted Array Given a sorted array that ... read more
asked in Cisco
Q4. Covid Vaccination Distribution Problem As the Government ramps up ... read more
asked in Sigmoid
Q5. K-th Element of Two Sorted Arrays You are provided with two sorte ... read more

Data Engineer Jobs at Persistent Systems

View all

Interview questions from similar companies

Interview experience
3
Average
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
No response

I applied via Naukri.com and was interviewed in Oct 2024. There were 2 interview rounds.

Round 1 - Technical 

(7 Questions)

  • Q1. How do you optimize SQL queries?
  • Ans. 

    Optimizing SQL queries involves using indexes, avoiding unnecessary joins, and optimizing the query structure.

    • Use indexes on columns frequently used in WHERE clauses

    • Avoid using SELECT * and only retrieve necessary columns

    • Optimize joins by using INNER JOIN instead of OUTER JOIN when possible

    • Use EXPLAIN to analyze query performance and make necessary adjustments

  • Answered by AI
  • Q2. How do you do performance optimization in Spark. Tell how you did it in you project.
  • Ans. 

    Performance optimization in Spark involves tuning configurations, optimizing code, and utilizing caching.

    • Tune Spark configurations such as executor memory, number of executors, and shuffle partitions.

    • Optimize code by reducing unnecessary shuffles, using efficient transformations, and avoiding unnecessary data movements.

    • Utilize caching to store intermediate results in memory and avoid recomputation.

    • Example: In my projec...

  • Answered by AI
  • Q3. What is SparkContext and SparkSession?
  • Ans. 

    SparkContext is the main entry point for Spark functionality, while SparkSession is the entry point for Spark SQL.

    • SparkContext is the entry point for low-level API functionality in Spark.

    • SparkSession is the entry point for Spark SQL functionality.

    • SparkContext is used to create RDDs (Resilient Distributed Datasets) in Spark.

    • SparkSession provides a unified entry point for reading data from various sources and performing

  • Answered by AI
  • Q4. When a spark job is submitted, what happens at backend. Explain the flow.
  • Ans. 

    When a spark job is submitted, various steps are executed at the backend to process the job.

    • The job is submitted to the Spark driver program.

    • The driver program communicates with the cluster manager to request resources.

    • The cluster manager allocates resources (CPU, memory) to the job.

    • The driver program creates DAG (Directed Acyclic Graph) of the job stages and tasks.

    • Tasks are then scheduled and executed on worker nodes ...

  • Answered by AI
  • Q5. Calculate second highest salary using SQL as well as pyspark.
  • Ans. 

    Calculate second highest salary using SQL and pyspark

    • Use SQL query with ORDER BY and LIMIT to get the second highest salary

    • In pyspark, use orderBy() and take() functions to achieve the same result

  • Answered by AI
  • Q6. 2 types of modes for Spark architecture ?
  • Ans. 

    The two types of modes for Spark architecture are standalone mode and cluster mode.

    • Standalone mode: Spark runs on a single machine with a single JVM and is suitable for development and testing.

    • Cluster mode: Spark runs on a cluster of machines managed by a cluster manager like YARN or Mesos for production workloads.

  • Answered by AI
  • Q7. If you want very less latency - which is better standalone or client mode?
  • Ans. 

    Client mode is better for very less latency due to direct communication with the cluster.

    • Client mode allows direct communication with the cluster, reducing latency.

    • Standalone mode requires an additional layer of communication, increasing latency.

    • Client mode is preferred for real-time applications where low latency is crucial.

  • Answered by AI
Round 2 - Technical 

(2 Questions)

  • Q1. Scenario based. Write SQL and pyspark code for a dataset.
  • Q2. If you have to find latest record based on latest timestamp in a table for a particular customer(table is having history) , how will you do it. Self join and nested query will be expensive. Optimized query...

Interview Preparation Tips

Topics to prepare for LTIMindtree Data Engineer interview:
  • SQL
  • pyspark
  • ETL
Interview preparation tips for other job seekers - L2 was scheduled next day to L1 so the process is fast. Brush up your practical knowledge more.

Skills evaluated in this interview

Interview experience
3
Average
Difficulty level
-
Process Duration
-
Result
No response
Round 1 - Technical 

(4 Questions)

  • Q1. What is the architecture of Apache Spark?
  • Ans. 

    Apache Spark architecture includes a cluster manager, worker nodes, and driver program.

    • Apache Spark architecture consists of a cluster manager, which allocates resources and schedules tasks.

    • Worker nodes execute tasks and store data in memory or disk.

    • Driver program coordinates tasks and communicates with the cluster manager.

    • Spark applications run as independent sets of processes on a cluster, coordinated by the SparkCon...

  • Answered by AI
  • Q2. What is the difference between the reduceBy and groupBy transformations in Apache Spark?
  • Ans. 

    reduceBy is used to aggregate data based on key, while groupBy is used to group data based on key.

    • reduceBy is a transformation that combines the values of each key using an associative function and a neutral 'zero value'.

    • groupBy is a transformation that groups the data based on a key and returns a grouped data set.

    • reduceBy is more efficient for aggregating data as it reduces the data before shuffling, while groupBy shu...

  • Answered by AI
  • Q3. What is the difference between RDD (Resilient Distributed Datasets) and DataFrame in Apache Spark?
  • Ans. 

    RDD is a low-level abstraction representing a distributed collection of objects, while DataFrame is a higher-level abstraction representing a distributed collection of data organized into named columns.

    • RDD is more suitable for unstructured data and low-level transformations, while DataFrame is more suitable for structured data and high-level abstractions.

    • DataFrames provide optimizations like query optimization and code...

  • Answered by AI
  • Q4. What are the different modes of execution in Apache Spark?
  • Ans. 

    The different modes of execution in Apache Spark include local mode, standalone mode, YARN mode, and Mesos mode.

    • Local mode: Spark runs on a single machine with one executor.

    • Standalone mode: Spark runs on a cluster managed by a standalone cluster manager.

    • YARN mode: Spark runs on a Hadoop cluster using YARN as the resource manager.

    • Mesos mode: Spark runs on a Mesos cluster with Mesos as the resource manager.

  • Answered by AI
Interview experience
3
Average
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Not Selected

I applied via LinkedIn and was interviewed in Nov 2024. There was 1 interview round.

Round 1 - Technical 

(2 Questions)

  • Q1. How to Ensure Data loss in ETL pipeline
  • Q2. Why to spin up Dataproc cluster when there is serverless batch job exists.
Interview experience
3
Average
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Not Selected

I applied via Naukri.com and was interviewed in Nov 2024. There was 1 interview round.

Round 1 - Technical 

(2 Questions)

  • Q1. Spark Architecture
  • Q2. Cache vs persist, lazy evaluation
Interview experience
4
Good
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Technical 

(2 Questions)

  • Q1. What is Databricks, Internals of it, Optimization technique
  • Ans. 

    Databricks is a unified analytics platform that provides a collaborative environment for data scientists, engineers, and analysts.

    • Databricks is built on top of Apache Spark, providing a unified platform for data engineering, data science, and business analytics.

    • Internals of Databricks include a cluster manager, job scheduler, and workspace for collaboration.

    • Optimization techniques in Databricks include query optimizati...

  • Answered by AI
  • Q2. SQL questions of joins,group by
Round 2 - Technical 

(2 Questions)

  • Q1. Scenario based azure data factory questions
  • Q2. Project structres, pyspark dataframes related

Interview Preparation Tips

Interview preparation tips for other job seekers - Juat practice fundamentals

Skills evaluated in this interview

Interview experience
4
Good
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Technical 

(1 Question)

  • Q1. Python SQL question
Round 2 - Technical 

(1 Question)

  • Q1. More on Project side
Round 3 - HR 

(1 Question)

  • Q1. Salary Discussion
Interview experience
1
Bad
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Technical 

(1 Question)

  • Q1. Data engineer roles and resposibilities

Interview Preparation Tips

Interview preparation tips for other job seekers - Don't Go, worst management service and lots of office politics.
Interview experience
4
Good
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Technical 

(2 Questions)

  • Q1. Python list comprehension
  • Q2. SQL- product with second highest total price
  • Ans. 

    Use SQL query to find the product with the second highest total price.

    • Use the ORDER BY clause to sort the products by total price in descending order

    • Use the LIMIT clause to select the second row after sorting

  • Answered by AI

Skills evaluated in this interview

Persistent Systems Interview FAQs

How many rounds are there in Persistent Systems Data Engineer interview?
Persistent Systems interview process usually has 1-2 rounds. The most common rounds in the Persistent Systems interview process are Technical and One-on-one Round.
How to prepare for Persistent Systems Data Engineer interview?
Go through your CV in detail and study all the technologies mentioned in your CV. Prepare at least two technologies or languages in depth if you are appearing for a technical interview at Persistent Systems. The most common topics and skills that interviewers at Persistent Systems expect are Python, Java, SQL, NoSQL and Spark.
What are the top questions asked in Persistent Systems Data Engineer interview?

Some of the top questions asked at the Persistent Systems Data Engineer interview -

  1. Aws sevices you have worked...read more
  2. lazy evaluation narrow wide tranform data brick utlity functi...read more
  3. Question on python list and diction...read more

Tell us how to improve this page.

Persistent Systems Data Engineer Interview Process

based on 5 interviews

Interview experience

4.2
  
Good
View more
Join Persistent Systems See Beyond, Rise Above
Persistent Systems Data Engineer Salary
based on 160 salaries
₹4.6 L/yr - ₹19.2 L/yr
9% more than the average Data Engineer Salary in India
View more details

Persistent Systems Data Engineer Reviews and Ratings

based on 14 reviews

3.6/5

Rating in categories

3.7

Skill development

3.7

Work-life balance

3.4

Salary

3.1

Job security

3.4

Company culture

3.2

Promotions

3.3

Work satisfaction

Explore 14 Reviews and Ratings
Data Engineer

Mumbai,

Hyderabad / Secunderabad

+1

5-10 Yrs

Not Disclosed

Explore more jobs
Software Engineer
4.4k salaries
unlock blur

₹2.5 L/yr - ₹9.5 L/yr

Senior Software Engineer
3.9k salaries
unlock blur

₹5 L/yr - ₹15.4 L/yr

Lead Engineer
3.1k salaries
unlock blur

₹10 L/yr - ₹30 L/yr

Lead Software Engineer
3k salaries
unlock blur

₹7.2 L/yr - ₹18.4 L/yr

Project Lead
1.9k salaries
unlock blur

₹13.2 L/yr - ₹38.8 L/yr

Explore more salaries
Compare Persistent Systems with

Cognizant

3.8
Compare

TCS

3.7
Compare

IBM

4.0
Compare

Wipro

3.7
Compare
Did you find this page helpful?
Yes No
write
Share an Interview