Upload Button Icon Add office photos
Premium Employer

i

This company page is being actively managed by NTT Data Team. If you also belong to the team, you can get access from here

NTT Data

Compare button icon Compare button icon Compare
3.9

based on 3.8k Reviews

Filter interviews by

NTT Data Senior Data Engineer Interview Questions and Answers

Updated 8 Jan 2025

NTT Data Senior Data Engineer Interview Experiences

1 interview found

Interview experience
4
Good
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Technical 

(1 Question)

  • Q1. Adf,etl,python,adb

Interview questions from similar companies

Interview experience
3
Average
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Aptitude Test 

The aptitude test lasts 30 minutes and focuses on topics relevant to data engineering, including Spark, SQL, Azure, and PySpark.

Round 2 - Coding Test 

The coding test is a one-hour examination on PySpark.

Round 3 - Technical 

(3 Questions)

  • Q1. What is the difference between Cache() and Persist()?
  • Q2. What does the purpose of the Spark Submit command in Apache Spark?
  • Q3. What are window functions in SQL?
Round 4 - HR 

(2 Questions)

  • Q1. Could you provide more details about the daily responsibilities associated with this role?
  • Q2. How would you describe your work culture?
Interview experience
5
Excellent
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Selected Selected

I applied via Naukri.com and was interviewed in Nov 2024. There were 2 interview rounds.

Round 1 - Technical 

(2 Questions)

  • Q1. All questions based on databricks and some pyspark , python , SQL only.
  • Q2. Learn windows function implementation in databricks note book.
Round 2 - HR 

(1 Question)

  • Q1. This round about your salary discussion around.

Interview Preparation Tips

Topics to prepare for Accenture Senior Data Engineer interview:
  • Python
  • Pyspark
  • SQL
  • Databricks
Interview preparation tips for other job seekers - Please prepare for pyspark, python , SQL , databricks for practice to switch your job to big data engineer
Interview experience
5
Excellent
Difficulty level
-
Process Duration
-
Result
-

I was interviewed in Dec 2024.

Round 1 - Technical 

(2 Questions)

  • Q1. Windows function-related questions.
  • Q2. Join Related Questions.
Round 2 - Technical 

(2 Questions)

  • Q1. Join related Questions
  • Q2. Subqueries related queations.
Interview experience
5
Excellent
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Selected Selected

I was interviewed in Aug 2024.

Round 1 - Coding Test 

Python and sql tasks

Round 2 - Technical 

(2 Questions)

  • Q1. Project related questions
  • Q2. Coding questions on pyspark windows
Round 3 - One-on-one 

(1 Question)

  • Q1. Managerial discussions, mostly in and around the previous projects
Interview experience
4
Good
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
No response

I was interviewed in Sep 2024.

Round 1 - Technical 

(3 Questions)

  • Q1. Coding in pyspark
  • Ans. 

    Pyspark is a Python library for big data processing using Spark framework.

    • Pyspark is used for processing large datasets in parallel.

    • It provides APIs for data manipulation, querying, and analysis.

    • Example: Using pyspark to read a CSV file and perform data transformations.

  • Answered by AI
  • Q2. Databricks optimisation technique
  • Ans. 

    Databricks optimisation techniques improve performance and efficiency of data processing on the Databricks platform.

    • Use cluster sizing and autoscaling to optimize resource allocation based on workload

    • Leverage Databricks Delta for optimized data storage and processing

    • Utilize caching and persisting data to reduce computation time

    • Optimize queries by using appropriate indexing and partitioning strategies

  • Answered by AI
  • Q3. Aqe details in databricks
  • Ans. 

    Databricks is a unified data analytics platform that provides a collaborative environment for data engineers.

    • Databricks is built on top of Apache Spark and provides a workspace for data engineering tasks.

    • It allows for easy integration with various data sources and tools for data processing.

    • Databricks provides features like notebooks, clusters, and libraries for efficient data engineering workflows.

  • Answered by AI

Skills evaluated in this interview

Interview experience
4
Good
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Selected Selected

I applied via Naukri.com and was interviewed in Jun 2024. There were 3 interview rounds.

Round 1 - Technical 

(2 Questions)

  • Q1. More questions on spark and java
  • Q2. More questions on sql
Round 2 - Technical 

(2 Questions)

  • Q1. Over all project architecture
  • Q2. Sample data and its transformations
  • Ans. 

    Sample data and its transformations

    • Sample data can be in the form of CSV, JSON, or database tables

    • Transformations include cleaning, filtering, aggregating, and joining data

    • Examples: converting date formats, removing duplicates, calculating averages

  • Answered by AI
Round 3 - HR 

(1 Question)

  • Q1. Why are u leaving current job
  • Ans. 

    Seeking new challenges and opportunities for growth in a more dynamic environment.

    • Looking for new challenges and opportunities for growth

    • Seeking a more dynamic work environment

    • Interested in expanding skill set and knowledge

    • Want to work on more innovative projects

  • Answered by AI
Interview experience
5
Excellent
Difficulty level
Hard
Process Duration
2-4 weeks
Result
-

I applied via Naukri.com and was interviewed in Sep 2024. There was 1 interview round.

Round 1 - One-on-one 

(2 Questions)

  • Q1. What is scd type 2?
  • Ans. 

    SCD type 2 is a method used in data warehousing to track historical changes by creating a new record for each change.

    • SCD type 2 stands for Slowly Changing Dimension type 2

    • It involves creating a new record in the dimension table whenever there is a change in the data

    • The old record is marked as inactive and the new record is marked as current

    • It allows for historical tracking of changes in data over time

    • Example: If a cust...

  • Answered by AI
  • Q2. Pyspark question read CSV from folder and add column in each csv file and write it to different location.

Skills evaluated in this interview

Interview experience
3
Average
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Technical 

(4 Questions)

  • Q1. Explain about yourself.
  • Ans. 

    I am a Senior Data Engineer with 5+ years of experience in designing and implementing data pipelines for large-scale projects.

    • Experienced in ETL processes and data warehousing

    • Proficient in programming languages like Python, SQL, and Java

    • Skilled in working with big data technologies such as Hadoop, Spark, and Kafka

    • Strong understanding of data modeling and database management

    • Excellent problem-solving and communication sk

  • Answered by AI
  • Q2. Explain about your current project.
  • Ans. 

    Developing a real-time data processing system for analyzing customer behavior on e-commerce platform.

    • Utilizing Apache Kafka for real-time data streaming

    • Implementing Spark for data processing and analysis

    • Creating machine learning models for customer segmentation

    • Integrating with Elasticsearch for data indexing and search functionality

  • Answered by AI
  • Q3. Big data concept
  • Q4. Spark related questions

Interview Preparation Tips

Interview preparation tips for other job seekers - Explore more technical and project process.
Interview experience
4
Good
Difficulty level
Moderate
Process Duration
2-4 weeks
Result
Not Selected

I applied via Naukri.com and was interviewed in Sep 2024. There was 1 interview round.

Round 1 - Coding Test 

Spark Optimization, Transformation, DLT, DL, Data Governance
Python
SQL

Interview Preparation Tips

Interview preparation tips for other job seekers - Ingestion, Integration, Spark, Optimization, Python, SQL, Data Warehouse

NTT Data Interview FAQs

How many rounds are there in NTT Data Senior Data Engineer interview?
NTT Data interview process usually has 1 rounds. The most common rounds in the NTT Data interview process are Technical.

Tell us how to improve this page.

NTT Data Senior Data Engineer Interview Process

based on 1 interview

Interview experience

4
  
Good
View more
NTT Data Senior Data Engineer Salary
based on 14 salaries
₹9.1 L/yr - ₹25 L/yr
19% less than the average Senior Data Engineer Salary in India
View more details

NTT Data Senior Data Engineer Reviews and Ratings

based on 1 review

3.0/5

Rating in categories

3.0

Skill development

2.0

Work-life balance

3.0

Salary

4.0

Job security

3.0

Company culture

3.0

Promotions

3.0

Work satisfaction

Explore 1 Review and Rating
Software Engineer
935 salaries
unlock blur

₹2.8 L/yr - ₹11 L/yr

Senior Associate
774 salaries
unlock blur

₹1.2 L/yr - ₹7.5 L/yr

Network Engineer
672 salaries
unlock blur

₹1.8 L/yr - ₹10 L/yr

Software Developer
641 salaries
unlock blur

₹3.1 L/yr - ₹12 L/yr

Senior Software Engineer
574 salaries
unlock blur

₹6.5 L/yr - ₹26 L/yr

Explore more salaries
Compare NTT Data with

Tata Communications

4.0
Compare

Bharti Airtel

4.0
Compare

Reliance Communications

4.0
Compare

Vodafone Idea

4.1
Compare
Did you find this page helpful?
Yes No
write
Share an Interview