Upload Button Icon Add office photos
Engaged Employer

i

This company page is being actively managed by Nagarro Team. If you also belong to the team, you can get access from here

Nagarro Verified Tick

Compare button icon Compare button icon Compare
4.0

based on 4k Reviews

Filter interviews by

Nagarro Senior Data Engineer Interview Questions and Answers

Updated 26 May 2022

Nagarro Senior Data Engineer Interview Experiences

2 interviews found

I applied via Recruitment Consulltant and was interviewed in Nov 2021. There were 4 interview rounds.

Round 1 - Aptitude Test 

Normal aptitude test

Round 2 - Technical 

(1 Question)

  • Q1. Technical discussion on big data and some programing problem
Round 3 - Technical 

(1 Question)

  • Q1. 2nd technical discussion based on project, big data and python
Round 4 - HR 

(1 Question)

  • Q1. Normal HR discussion and salary negotiation

Interview Preparation Tips

Interview preparation tips for other job seekers - Brush up your core skill and basic programming knowledge

I applied via Recruitment Consultant and was interviewed in Jun 2021. There were 4 interview rounds.

Interview Questionnaire 

2 Questions

  • Q1. Data pipelines set up in Azure data factory, Any experience in programming language like python, Java.
  • Q2. Power bi and data modeling experience

Interview Preparation Tips

Interview preparation tips for other job seekers - Try to explain your current project architecture to showcase your knowledge. Mostly people are looking for a end to end developer, with experience in data modeling, data orchestration and reporting.
Any two areas would work perfectly fine.

Senior Data Engineer Interview Questions Asked at Other Companies

asked in 7 Eleven
Q1. Write a query to get the customer with the highest total order va ... read more
asked in 7 Eleven
Q2. There are 10 million records in the table and the schema does not ... read more
asked in 7 Eleven
Q3. How do you handle data pipeline when the schema information keeps ... read more
asked in 7 Eleven
Q4. Difference between Parquet and ORC file. Why industry uses parque ... read more
asked in 7 Eleven
Q5. What is Normalisation and Denormalisation? When do we use them? G ... read more

Interview questions from similar companies

Interview experience
3
Average
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Aptitude Test 

The aptitude test lasts 30 minutes and focuses on topics relevant to data engineering, including Spark, SQL, Azure, and PySpark.

Round 2 - Coding Test 

The coding test is a one-hour examination on PySpark.

Round 3 - Technical 

(3 Questions)

  • Q1. What is the difference between Cache() and Persist()?
  • Q2. What does the purpose of the Spark Submit command in Apache Spark?
  • Q3. What are window functions in SQL?
Round 4 - HR 

(2 Questions)

  • Q1. Could you provide more details about the daily responsibilities associated with this role?
  • Q2. How would you describe your work culture?
Interview experience
5
Excellent
Difficulty level
-
Process Duration
-
Result
-

I was interviewed in Dec 2024.

Round 1 - Technical 

(2 Questions)

  • Q1. Windows function-related questions.
  • Q2. Join Related Questions.
Round 2 - Technical 

(2 Questions)

  • Q1. Join related Questions
  • Q2. Subqueries related queations.
Interview experience
3
Average
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Not Selected

I applied via Naukri.com and was interviewed in Nov 2024. There was 1 interview round.

Round 1 - Technical 

(2 Questions)

  • Q1. How do you utilize the enhanced optimization option in AWS Glue?
  • Ans. 

    Enhanced optimization in AWS Glue improves job performance by automatically adjusting resources based on workload

    • Enhanced optimization in AWS Glue automatically adjusts resources like DPUs based on workload

    • It helps improve job performance by optimizing resource allocation

    • Users can enable enhanced optimization in AWS Glue job settings

  • Answered by AI
  • Q2. What are the best practices for optimizing querying in Amazon Redshift?
  • Ans. 

    Optimizing querying in Amazon Redshift involves proper table design, distribution keys, sort keys, and query optimization techniques.

    • Use appropriate distribution keys to evenly distribute data across nodes for parallel processing.

    • Utilize sort keys to physically order data on disk, reducing the need for sorting during queries.

    • Avoid using SELECT * and instead specify only the columns needed to reduce data transfer.

    • Use AN...

  • Answered by AI
Interview experience
4
Good
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Technical 

(1 Question)

  • Q1. Adf,etl,python,adb
Interview experience
4
Good
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Selected Selected

I applied via Naukri.com and was interviewed in Jun 2024. There were 3 interview rounds.

Round 1 - Technical 

(2 Questions)

  • Q1. More questions on spark and java
  • Q2. More questions on sql
Round 2 - Technical 

(2 Questions)

  • Q1. Over all project architecture
  • Q2. Sample data and its transformations
  • Ans. 

    Sample data and its transformations

    • Sample data can be in the form of CSV, JSON, or database tables

    • Transformations include cleaning, filtering, aggregating, and joining data

    • Examples: converting date formats, removing duplicates, calculating averages

  • Answered by AI
Round 3 - HR 

(1 Question)

  • Q1. Why are u leaving current job
  • Ans. 

    Seeking new challenges and opportunities for growth in a more dynamic environment.

    • Looking for new challenges and opportunities for growth

    • Seeking a more dynamic work environment

    • Interested in expanding skill set and knowledge

    • Want to work on more innovative projects

  • Answered by AI
Interview experience
5
Excellent
Difficulty level
Hard
Process Duration
2-4 weeks
Result
-

I applied via Naukri.com and was interviewed in Sep 2024. There was 1 interview round.

Round 1 - One-on-one 

(2 Questions)

  • Q1. What is scd type 2?
  • Ans. 

    SCD type 2 is a method used in data warehousing to track historical changes by creating a new record for each change.

    • SCD type 2 stands for Slowly Changing Dimension type 2

    • It involves creating a new record in the dimension table whenever there is a change in the data

    • The old record is marked as inactive and the new record is marked as current

    • It allows for historical tracking of changes in data over time

    • Example: If a cust...

  • Answered by AI
  • Q2. Pyspark question read CSV from folder and add column in each csv file and write it to different location.

Skills evaluated in this interview

Interview experience
3
Average
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Not Selected

I applied via Job Portal and was interviewed in Jul 2024. There was 1 interview round.

Round 1 - HR 

(2 Questions)

  • Q1. Tell about experience
  • Ans. 

    I have over 5 years of experience in data engineering, working with large datasets and implementing data pipelines.

    • Developed and maintained ETL processes to extract, transform, and load data from various sources

    • Optimized database performance and implemented data quality checks

    • Worked with cross-functional teams to design and implement data solutions

    • Utilized tools such as Apache Spark, Hadoop, and SQL for data processing

    • ...

  • Answered by AI
  • Q2. What would you do when you are given a new task to do
  • Ans. 

    I would start by understanding the requirements, breaking down the task into smaller steps, researching if needed, and then creating a plan to execute the task efficiently.

    • Understand the requirements of the task

    • Break down the task into smaller steps

    • Research if needed to gather necessary information

    • Create a plan to execute the task efficiently

    • Communicate with stakeholders for clarification or updates

    • Regularly track prog

  • Answered by AI
Interview experience
4
Good
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Not Selected

I applied via Company Website and was interviewed in Sep 2024. There were 2 interview rounds.

Round 1 - Coding Test 

Platform - Hackerank
Duration - 2 Hours
Topics - Spark and SQL

Round 2 - Technical 

(3 Questions)

  • Q1. What are the common file formats used in data storages? Which one is best for compression?
  • Ans. 

    Common file formats used in data storages include CSV, JSON, Parquet, Avro, and ORC. Parquet is best for compression.

    • CSV (Comma-Separated Values) - simple and widely used, but not efficient for large datasets

    • JSON (JavaScript Object Notation) - human-readable and easy to parse, but can be inefficient for storage

    • Parquet - columnar storage format that is highly efficient for compression and query performance

    • Avro - efficie...

  • Answered by AI
  • Q2. SQL Problem - Given the empoyee attendance table, write a query to print the employees who is abscent for more than cosecutive 10 days in their tenure.
  • Q3. Given the list of words, write the Python program to print the most repeating substring out of all words.
  • Ans. 

    Python program to find the most repeating substring in a list of words.

    • Iterate through each word in the list

    • Generate all possible substrings for each word

    • Count the occurrences of each substring using a dictionary

    • Find the substring with the highest count

  • Answered by AI

Interview Preparation Tips

Interview preparation tips for other job seekers - Prepare well in SQL, Spark and Python coding problems.

Skills evaluated in this interview

Nagarro Interview FAQs

How many rounds are there in Nagarro Senior Data Engineer interview?
Nagarro interview process usually has 4 rounds. The most common rounds in the Nagarro interview process are Technical, HR and Aptitude Test.
What are the top questions asked in Nagarro Senior Data Engineer interview?

Some of the top questions asked at the Nagarro Senior Data Engineer interview -

  1. Data pipelines set up in Azure data factory, Any experience in programming lang...read more
  2. Technical discussion on big data and some programing prob...read more
  3. 2nd technical discussion based on project, big data and pyt...read more

Tell us how to improve this page.

People are getting interviews through

based on 2 Nagarro interviews
Recruitment Consultant
50%
50% candidates got the interview through other sources.
Moderate Confidence
?
Moderate Confidence means the data is based on a sufficient number of responses received from the candidates
Nagarro Senior Data Engineer Salary
based on 47 salaries
₹10 L/yr - ₹38.2 L/yr
17% more than the average Senior Data Engineer Salary in India
View more details

Nagarro Senior Data Engineer Reviews and Ratings

based on 3 reviews

4.0/5

Rating in categories

4.0

Skill development

4.0

Work-Life balance

4.5

Salary & Benefits

4.5

Job Security

5.0

Company culture

4.0

Promotions/Appraisal

4.0

Work Satisfaction

Explore 3 Reviews and Ratings
Associate Staff Engineer
2.9k salaries
unlock blur

₹10 L/yr - ₹35.2 L/yr

Staff Engineer
2.8k salaries
unlock blur

₹14.5 L/yr - ₹43.2 L/yr

Senior Engineer
2.3k salaries
unlock blur

₹6.3 L/yr - ₹25 L/yr

Senior Software Engineer
1.1k salaries
unlock blur

₹7.5 L/yr - ₹31 L/yr

Software Engineer
938 salaries
unlock blur

₹3 L/yr - ₹12.2 L/yr

Explore more salaries
Compare Nagarro with

Deloitte

3.8
Compare

Cognizant

3.8
Compare

TCS

3.7
Compare

Accenture

3.9
Compare

Calculate your in-hand salary

Confused about how your in-hand salary is calculated? Enter your annual salary (CTC) and get your in-hand salary
Did you find this page helpful?
Yes No
write
Share an Interview