Upload Button Icon Add office photos
Engaged Employer

i

This company page is being actively managed by TCS Team. If you also belong to the team, you can get access from here

TCS Verified Tick

Compare button icon Compare button icon Compare

Filter interviews by

TCS Senior Data Engineer Interview Questions and Answers

Updated 8 Apr 2025

24 Interview questions

A Senior Data Engineer was asked 2mo ago
Q. In BigQuery, where do NULL values go when using PARTITION BY?
Ans. 

In BigQuery, null values are treated as a separate partition in partitioned tables.

  • Null values are grouped together in their own partition when using partitioned tables.

  • For example, if you partition by a date column, all rows with null dates will go into a 'null' partition.

  • This allows for efficient querying, as queries can skip the null partition if not needed.

  • In a partitioned table, null values do not affect the ...

A Senior Data Engineer was asked 2mo ago
Q. Can we perform clustering after partitioning the table?
Ans. 

Clustering can be performed after partitioning a table, enhancing data organization and query performance.

  • Partitioning divides a table into smaller, manageable pieces based on a key (e.g., date, region).

  • Clustering organizes data within each partition based on specific attributes (e.g., customer ID, product type).

  • Example: A sales table can be partitioned by year and clustered by region to optimize queries for speci...

Senior Data Engineer Interview Questions Asked at Other Companies

asked in 7 Eleven
Q1. Write a query to get the customer with the highest total order va ... read more
asked in 7 Eleven
Q2. There are 10 million records in the table and the schema does not ... read more
asked in KFintech
Q3. Given infinite coins of some currency of denominations : 1,2,5,10 ... read more
asked in 7 Eleven
Q4. How do you handle data pipelines when the schema information keep ... read more
asked in 7 Eleven
Q5. Difference between Parquet and ORC file. Why industry uses parque ... read more
A Senior Data Engineer was asked 6mo ago
Q. Why does Spark work well with Parquet files?
Ans. 

Spark works well with Parquet files due to its columnar storage format, efficient compression, and ability to push down filters.

  • Parquet files are columnar storage format, which aligns well with Spark's processing model of working on columns rather than rows.

  • Parquet files support efficient compression, reducing storage space and improving read performance in Spark.

  • Spark can push down filters to Parquet files, allow...

A Senior Data Engineer was asked 6mo ago
Q. What are constructors in Python?
Ans. 

Constructors in Python are special methods used for initializing objects. They are called automatically when a new instance of a class is created.

  • Constructors are defined using the __init__() method in a class.

  • They are used to initialize instance variables of a class.

  • Example: class Person: def __init__(self, name, age): self.name = name self.age = age person1 = Person('Alice', 30)

What people are saying about TCS

View All
a digital marketer
3d
Do you think they're gonna work on employees' work-life balance, OR for just publicity?
Infosys, Infosys, TCS, Genpact Revise Workplace Policies Infosys is sending a warning mail, if an employee overshoots the daily limit while working remotely, the system triggers a notification Genpact introduced a new policy to log in before 11 am But will these companies really change, or is it just a show to mask their issues?
FeedCard Image
Got a question about TCS?
Ask anonymously on communities.
A Senior Data Engineer was asked 6mo ago
Q. What is the SQL query to group by employee ID in order to combine the first name and last name with a space?
Ans. 

SQL query to group by employee ID and combine first name and last name with a space

  • Use the GROUP BY clause to group by employee ID

  • Use the CONCAT function to combine first name and last name with a space

  • Select employee ID, CONCAT(first_name, ' ', last_name) AS full_name

A Senior Data Engineer was asked 8mo ago
Q. What is SQL? Please explain.
Ans. 

SQL is a programming language used for managing and manipulating relational databases.

  • SQL stands for Structured Query Language

  • It is used to communicate with databases to perform tasks such as querying data, updating data, and creating tables

  • Common SQL commands include SELECT, INSERT, UPDATE, DELETE

  • Example: SELECT * FROM employees WHERE department = 'IT'

A Senior Data Engineer was asked 8mo ago
Q. What is Python? Please explain.
Ans. 

Python is a high-level programming language known for its simplicity and readability.

  • Python is an interpreted language, meaning code is executed line by line.

  • It is dynamically typed, allowing for flexibility in variable types.

  • Python is popular for web development, data analysis, artificial intelligence, and more.

  • Example: print('Hello, World!') is a simple Python program to display text.

Are these interview questions helpful?
A Senior Data Engineer was asked 8mo ago
Q. How do you deploy a Spark application?
Ans. 

Spark applications can be deployed using various methods like standalone mode, YARN, Mesos, or Kubernetes.

  • Deploy Spark application in standalone mode by submitting the application using spark-submit command

  • Deploy Spark application on YARN by setting the master to yarn and submitting the application to the YARN ResourceManager

  • Deploy Spark application on Mesos by setting the master to mesos and submitting the applic...

A Senior Data Engineer was asked 8mo ago
Q. How do you decide on the Spark configuration for a job?
Ans. 

Spark configuration for a job is decided based on factors like data size, cluster resources, and job requirements.

  • Consider the size of the data being processed to determine the number of partitions and memory requirements.

  • Evaluate the available cluster resources such as CPU cores, memory, and storage to optimize performance.

  • Adjust parameters like executor memory, executor cores, and driver memory based on the comp...

A Senior Data Engineer was asked 8mo ago
Q. What are the types of SCD?
Ans. 

Types of SCD include Type 1, Type 2, and Type 3.

  • Type 1 SCD: Overwrites old data with new data, no history is maintained.

  • Type 2 SCD: Maintains historical data by creating new records for changes.

  • Type 3 SCD: Creates separate columns to store historical and current data.

  • Examples: Type 1 - Employee address updates overwrite old address. Type 2 - Employee salary changes create new record with effective date. Type 3 - E...

TCS Senior Data Engineer Interview Experiences

12 interviews found

Interview experience
4
Good
Difficulty level
Easy
Process Duration
Less than 2 weeks
Result
No response

I applied via Naukri.com and was interviewed in Nov 2024. There was 1 interview round.

Round 1 - Technical 

(6 Questions)

  • Q1. Can you introduce yourself and describe your current project experience?
  • Ans. 

    I am a Senior Data Engineer with experience in building scalable data pipelines and optimizing data processing workflows.

    • Experience in designing and implementing ETL processes using tools like Apache Spark and Airflow

    • Proficient in working with large datasets and optimizing query performance

    • Strong background in data modeling and database design

    • Worked on projects involving real-time data processing and streaming analytic...

  • Answered by AI
  • Q2. Decorators in python
  • Ans. 

    Decorators in Python are functions that modify the behavior of other functions or methods.

    • Decorators are defined using the @decorator_name syntax before a function definition.

    • They can be used to add functionality to existing functions without modifying their code.

    • Decorators can be used for logging, timing, authentication, and more.

    • Example: @staticmethod decorator in Python is used to define a static method in a class.

  • Answered by AI
  • Q3. What is the SQL query to group by employee ID in order to combine the first name and last name with a space?
  • Ans. 

    SQL query to group by employee ID and combine first name and last name with a space

    • Use the GROUP BY clause to group by employee ID

    • Use the CONCAT function to combine first name and last name with a space

    • Select employee ID, CONCAT(first_name, ' ', last_name) AS full_name

  • Answered by AI
  • Q4. What are constructors in Python?
  • Ans. 

    Constructors in Python are special methods used for initializing objects. They are called automatically when a new instance of a class is created.

    • Constructors are defined using the __init__() method in a class.

    • They are used to initialize instance variables of a class.

    • Example: class Person: def __init__(self, name, age): self.name = name self.age = age person1 = Person('Alice', 30)

  • Answered by AI
  • Q5. Indexing in sql
  • Ans. 

    Indexing in SQL is a technique used to improve the performance of queries by creating a data structure that allows for faster retrieval of data.

    • Indexes are created on columns in a database table to speed up the retrieval of rows that match a certain condition in a WHERE clause.

    • Indexes can be created using CREATE INDEX statement in SQL.

    • Types of indexes include clustered indexes, non-clustered indexes, unique indexes, an...

  • Answered by AI
  • Q6. Why spark works well with parquet files?
  • Ans. 

    Spark works well with Parquet files due to its columnar storage format, efficient compression, and ability to push down filters.

    • Parquet files are columnar storage format, which aligns well with Spark's processing model of working on columns rather than rows.

    • Parquet files support efficient compression, reducing storage space and improving read performance in Spark.

    • Spark can push down filters to Parquet files, allowing f...

  • Answered by AI

Skills evaluated in this interview

Interview experience
5
Excellent
Difficulty level
Moderate
Process Duration
2-4 weeks
Result
Not Selected

I applied via Naukri.com and was interviewed in Aug 2024. There were 2 interview rounds.

Round 1 - Technical 

(8 Questions)

  • Q1. Tell me about your project.
  • Ans. 

    Developed a real-time data processing system for analyzing customer behavior on an e-commerce platform.

    • Used Apache Kafka for real-time data streaming

    • Implemented data pipelines using Apache Spark for processing large volumes of data

    • Designed and optimized data models in PostgreSQL for storing and querying customer data

  • Answered by AI
  • Q2. What are the Types of SCD?
  • Ans. 

    Types of SCD include Type 1, Type 2, and Type 3.

    • Type 1 SCD: Overwrites old data with new data, no history is maintained.

    • Type 2 SCD: Maintains historical data by creating new records for changes.

    • Type 3 SCD: Creates separate columns to store historical and current data.

    • Examples: Type 1 - Employee address updates overwrite old address. Type 2 - Employee salary changes create new record with effective date. Type 3 - Employ...

  • Answered by AI
  • Q3. Delta Lake vs Data Lake?
  • Ans. 

    Delta Lake is an open-source storage layer that brings ACID transactions to Apache Spark and big data workloads.

    • Delta Lake provides ACID transactions, schema enforcement, and data versioning on top of data lakes.

    • Data Lake is a centralized repository that allows you to store all your structured and unstructured data at any scale.

    • Delta Lake is optimized for big data workloads and is built on top of Apache Spark.

    • Data Lake...

  • Answered by AI
  • Q4. How to write a file in a delta table?
  • Ans. 

    To write a file in a delta table, you can use the Delta Lake API or Spark SQL commands.

    • Use Delta Lake API to write data to a delta table

    • Use Spark SQL commands like INSERT INTO to write data to a delta table

    • Ensure that the data being written is in the correct format and schema

  • Answered by AI
  • Q5. What are optimisation techniques used in the project?
  • Ans. 

    Optimisation techniques used in the project include indexing, query optimization, caching, and parallel processing.

    • Indexing: Creating indexes on frequently queried columns to improve search performance.

    • Query optimization: Rewriting queries to make them more efficient and reduce execution time.

    • Caching: Storing frequently accessed data in memory to reduce the need for repeated database queries.

    • Parallel processing: Distri...

  • Answered by AI
  • Q6. What is the difference between Tasks, and stages? About Spark UI?
  • Ans. 

    Tasks and stages are components of the execution plan in Spark UI.

    • Tasks are the smallest unit of work in Spark, representing a single operation on a partition of data.

    • Stages are groups of tasks that are executed together as part of a larger computation.

    • Tasks within a stage can be executed in parallel, while stages are executed sequentially.

    • Tasks are created based on the transformations and actions in the Spark applicat...

  • Answered by AI
  • Q7. Role of DAG ins aprk ?
  • Ans. 

    DAG (Directed Acyclic Graph) in Apache Spark is used to represent a series of data processing steps and their dependencies.

    • DAG in Spark helps optimize the execution of tasks by determining the order in which they should be executed based on dependencies.

    • It breaks down a Spark job into smaller tasks and organizes them in a way that minimizes unnecessary computations.

    • DAGs are created automatically by Spark when actions a...

  • Answered by AI
  • Q8. What transformation techniques you have used in your project?
  • Ans. 

    I have used various transformation techniques such as data cleaning, normalization, aggregation, and feature engineering in my projects.

    • Data cleaning to remove missing values and outliers

    • Normalization to scale numerical features

    • Aggregation to summarize data at different levels

    • Feature engineering to create new relevant features

  • Answered by AI
Round 2 - One-on-one 

(3 Questions)

  • Q1. About my project?
  • Q2. Some scenario-based questions related to project and work?
  • Q3. How do I keep myself updated with the latest technologies?
  • Ans. 

    I keep myself updated by regularly attending conferences, workshops, online courses, and reading industry blogs.

    • Attend conferences and workshops related to data engineering

    • Take online courses on platforms like Coursera, Udemy, or DataCamp

    • Read industry blogs and follow thought leaders on social media

    • Join online communities and forums to discuss latest technologies and trends

  • Answered by AI

Interview Preparation Tips

Topics to prepare for TCS Senior Data Engineer interview:
  • SQL
  • Spark
  • AWS
  • Project
Interview preparation tips for other job seekers - Keep the focus on your Spark fundaments, AWS Services and Solid Understanding of your project.

Skills evaluated in this interview

Interview experience
4
Good
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Selected Selected

I applied via Recruitment Consulltant and was interviewed in May 2024. There were 2 interview rounds.

Round 1 - Coding Test 

Questions related to GCP and cloud functions

Round 2 - Technical 

(6 Questions)

  • Q1. Introduce yourself, what role are you working as and what are your responsibilities.
  • Q2. Could you please explain GCP architecture?
  • Ans. 

    GCP architecture refers to the structure and components of Google Cloud Platform for building and managing applications and services.

    • GCP architecture is based on a global network of data centers that provide secure, scalable infrastructure for cloud services.

    • Key components include Compute Engine for virtual machines, Cloud Storage for object storage, and BigQuery for data analytics.

    • GCP architecture also includes networ...

  • Answered by AI
  • Q3. How can you optimize your queries for efficiency in BQ?
  • Ans. 

    Optimizing queries in BigQuery involves using partitioned tables, clustering, and optimizing joins.

    • Partition tables by date or another relevant column to reduce the amount of data scanned

    • Use clustering to group related rows together, reducing the amount of data scanned for queries

    • Avoid unnecessary joins and denormalize data where possible to reduce query complexity

  • Answered by AI
  • Q4. Do you have experience in Dataflow, Dataproc, cloud composer?
  • Ans. 

    Yes, I have experience in Dataflow, Dataproc, and cloud composer.

    • I have worked with Dataflow to process and analyze large datasets in real-time.

    • I have used Dataproc to create and manage Apache Spark and Hadoop clusters for big data processing.

    • I have experience with cloud composer for orchestrating workflows and managing data pipelines.

  • Answered by AI
  • Q5. What are the different joins in SQL? Please give an example to elaborate.
  • Ans. 

    Different types of joins in SQL include inner join, left join, right join, and full outer join.

    • Inner join: Returns rows when there is a match in both tables.

    • Left join: Returns all rows from the left table and the matched rows from the right table.

    • Right join: Returns all rows from the right table and the matched rows from the left table.

    • Full outer join: Returns rows when there is a match in either table.

    • Example: SELECT ...

  • Answered by AI
  • Q6. Explain cloud functions like cloud build, cloud run in GCP.
  • Ans. 

    Cloud functions like Cloud Build and Cloud Run in GCP are serverless computing services for building and running applications in the cloud.

    • Cloud Build is a service that executes your builds on Google Cloud Platform infrastructure. It automatically builds and tests your code in the cloud.

    • Cloud Run is a managed compute platform that enables you to run stateless containers that are invocable via HTTP requests. It automati...

  • Answered by AI

Interview Preparation Tips

Interview preparation tips for other job seekers - Be proactive and listen carefully. Also do not interrupt while the interviewer asks you follow up questions.
Also be clear and calm when you answer.
Salutations wherever necessary.

Skills evaluated in this interview

Interview experience
4
Good
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Technical 

(2 Questions)

  • Q1. How do you deploy spark application
  • Ans. 

    Spark applications can be deployed using various methods like standalone mode, YARN, Mesos, or Kubernetes.

    • Deploy Spark application in standalone mode by submitting the application using spark-submit command

    • Deploy Spark application on YARN by setting the master to yarn and submitting the application to the YARN ResourceManager

    • Deploy Spark application on Mesos by setting the master to mesos and submitting the application...

  • Answered by AI
  • Q2. How do you decide Spark configuration for a job
  • Ans. 

    Spark configuration for a job is decided based on factors like data size, cluster resources, and job requirements.

    • Consider the size of the data being processed to determine the number of partitions and memory requirements.

    • Evaluate the available cluster resources such as CPU cores, memory, and storage to optimize performance.

    • Adjust parameters like executor memory, executor cores, and driver memory based on the complexit...

  • Answered by AI

Skills evaluated in this interview

Interview experience
1
Bad
Difficulty level
Easy
Process Duration
Less than 2 weeks
Result
No response

I applied via Walk-in and was interviewed in Apr 2024. There were 2 interview rounds.

Round 1 - Technical 

(3 Questions)

  • Q1. No questions just asked to write SQL query and Python script. Not sure what they expect from it.
  • Q2. Write a SQL query
  • Ans. 

    SQL query to retrieve all employees from a table named 'employees'

    • Use SELECT * FROM employees;

    • Replace '*' with specific columns if needed, e.g. SELECT employee_id, name FROM employees;

  • Answered by AI
  • Q3. Write a Python program
  • Ans. 

    Python program to print 'Hello, World!'

    • Use the print() function in Python to display text on the screen

    • Enclose the text in single or double quotes to indicate a string

  • Answered by AI
Round 2 - One-on-one 

(1 Question)

  • Q1. Self intro and questions based on previous experiance

Skills evaluated in this interview

Interview experience
5
Excellent
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Selected Selected

I appeared for an interview in Oct 2024, where I was asked the following questions.

  • Q1. Where does null values goes in partition by in BigQuery?
  • Ans. 

    In BigQuery, null values are treated as a separate partition in partitioned tables.

    • Null values are grouped together in their own partition when using partitioned tables.

    • For example, if you partition by a date column, all rows with null dates will go into a 'null' partition.

    • This allows for efficient querying, as queries can skip the null partition if not needed.

    • In a partitioned table, null values do not affect the distr...

  • Answered by AI
  • Q2. Can we do clustering with partitioning the table?
  • Ans. 

    Clustering can be performed after partitioning a table, enhancing data organization and query performance.

    • Partitioning divides a table into smaller, manageable pieces based on a key (e.g., date, region).

    • Clustering organizes data within each partition based on specific attributes (e.g., customer ID, product type).

    • Example: A sales table can be partitioned by year and clustered by region to optimize queries for specific r...

  • Answered by AI
Interview experience
4
Good
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Coding Test 

Big data questins were asked gave scenario based questions

Very good response frok. Interviewer, he knows what he is expecting

Senior Data Engineer Interview Questions & Answers

user image Priyanshi Pundir

posted on 29 Sep 2024

Interview experience
5
Excellent
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Coding Test 

Code for printing duplicate numbers in a list.

Interview experience
3
Average
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
No response

I applied via Job Portal and was interviewed in Mar 2024. There were 2 interview rounds.

Round 1 - Technical 

(2 Questions)

  • Q1. Python Theoritical
  • Q2. Pyspark coding questions
Round 2 - HR 

(1 Question)

  • Q1. Salary and location
Interview experience
5
Excellent
Difficulty level
Moderate
Process Duration
2-4 weeks
Result
Selected Selected

I applied via Naukri.com and was interviewed before Oct 2023. There were 2 interview rounds.

Round 1 - Coding Test 

SQl, Python Data factory, data bricks, snowflake

Round 2 - Technical 

(2 Questions)

  • Q1. What is SQL and explain clearly
  • Ans. 

    SQL is a programming language used for managing and manipulating relational databases.

    • SQL stands for Structured Query Language

    • It is used to communicate with databases to perform tasks such as querying data, updating data, and creating tables

    • Common SQL commands include SELECT, INSERT, UPDATE, DELETE

    • Example: SELECT * FROM employees WHERE department = 'IT'

  • Answered by AI
  • Q2. What is python and explain clearly
  • Ans. 

    Python is a high-level programming language known for its simplicity and readability.

    • Python is an interpreted language, meaning code is executed line by line.

    • It is dynamically typed, allowing for flexibility in variable types.

    • Python is popular for web development, data analysis, artificial intelligence, and more.

    • Example: print('Hello, World!') is a simple Python program to display text.

  • Answered by AI

Skills evaluated in this interview

TCS Interview FAQs

How many rounds are there in TCS Senior Data Engineer interview?
TCS interview process usually has 1-2 rounds. The most common rounds in the TCS interview process are Technical, Coding Test and One-on-one Round.
How to prepare for TCS Senior Data Engineer interview?
Go through your CV in detail and study all the technologies mentioned in your CV. Prepare at least two technologies or languages in depth if you are appearing for a technical interview at TCS. The most common topics and skills that interviewers at TCS expect are ETL, Hive, Python, Spark and Hadoop.
What are the top questions asked in TCS Senior Data Engineer interview?

Some of the top questions asked at the TCS Senior Data Engineer interview -

  1. What is the difference between Tasks, and stages? About Spark ...read more
  2. What are optimisation techniques used in the proje...read more
  3. What is the SQL query to group by employee ID in order to combine the first nam...read more
How long is the TCS Senior Data Engineer interview process?

The duration of TCS Senior Data Engineer interview process can vary, but typically it takes about less than 2 weeks to complete.

Tell us how to improve this page.

Overall Interview Experience Rating

3.9/5

based on 12 interview experiences

Difficulty level

Easy 33%
Moderate 67%

Duration

Less than 2 weeks 67%
2-4 weeks 33%
View more
TCS Senior Data Engineer Salary
based on 422 salaries
₹14.5 L/yr - ₹25 L/yr
5% less than the average Senior Data Engineer Salary in India
View more details

TCS Senior Data Engineer Reviews and Ratings

based on 20 reviews

3.9/5

Rating in categories

3.6

Skill development

3.9

Work-life balance

3.0

Salary

4.8

Job security

3.6

Company culture

2.4

Promotions

3.4

Work satisfaction

Explore 20 Reviews and Ratings
System Engineer
1.1L salaries
unlock blur

₹3.9 L/yr - ₹8.3 L/yr

IT Analyst
65.5k salaries
unlock blur

₹7.7 L/yr - ₹12.7 L/yr

AST Consultant
53.6k salaries
unlock blur

₹12 L/yr - ₹20.6 L/yr

Assistant System Engineer
33.2k salaries
unlock blur

₹2.5 L/yr - ₹6.4 L/yr

Associate Consultant
33k salaries
unlock blur

₹16.2 L/yr - ₹28 L/yr

Explore more salaries
Compare TCS with

Amazon

4.0
Compare

Wipro

3.7
Compare

Infosys

3.6
Compare

Accenture

3.7
Compare
write
Share an Interview