i
NTT Data
Filter interviews by
I applied via Approached by Company and was interviewed before May 2023. There was 1 interview round.
I applied via Naukri.com and was interviewed in Nov 2024. There were 2 interview rounds.
The Aptitude Test session accesses mathematical and logical reasoning abilities
posted on 10 Jan 2025
I applied via Approached by Company and was interviewed in Dec 2024. There was 1 interview round.
posted on 31 Dec 2024
Apache Spark architecture includes a cluster manager, worker nodes, and driver program.
Apache Spark architecture consists of a cluster manager, which allocates resources and schedules tasks.
Worker nodes execute tasks and store data in memory or disk.
Driver program coordinates tasks and communicates with the cluster manager.
Spark applications run as independent sets of processes on a cluster, coordinated by the SparkCon...
reduceBy is used to aggregate data based on key, while groupBy is used to group data based on key.
reduceBy is a transformation that combines the values of each key using an associative function and a neutral 'zero value'.
groupBy is a transformation that groups the data based on a key and returns a grouped data set.
reduceBy is more efficient for aggregating data as it reduces the data before shuffling, while groupBy shu...
RDD is a low-level abstraction representing a distributed collection of objects, while DataFrame is a higher-level abstraction representing a distributed collection of data organized into named columns.
RDD is more suitable for unstructured data and low-level transformations, while DataFrame is more suitable for structured data and high-level abstractions.
DataFrames provide optimizations like query optimization and code...
The different modes of execution in Apache Spark include local mode, standalone mode, YARN mode, and Mesos mode.
Local mode: Spark runs on a single machine with one executor.
Standalone mode: Spark runs on a cluster managed by a standalone cluster manager.
YARN mode: Spark runs on a Hadoop cluster using YARN as the resource manager.
Mesos mode: Spark runs on a Mesos cluster with Mesos as the resource manager.
I applied via Recruitment Consulltant and was interviewed in Nov 2024. There were 2 interview rounds.
Different types of joins available in Databricks include inner join, outer join, left join, right join, and cross join.
Inner join: Returns only the rows that have matching values in both tables.
Outer join: Returns all rows when there is a match in either table.
Left join: Returns all rows from the left table and the matched rows from the right table.
Right join: Returns all rows from the right table and the matched rows ...
Implementing fault tolerance in a data pipeline involves redundancy, monitoring, and error handling.
Use redundant components to ensure continuous data flow
Implement monitoring tools to detect failures and bottlenecks
Set up automated alerts for immediate response to issues
Design error handling mechanisms to gracefully handle failures
Use checkpoints and retries to ensure data integrity
AutoLoader is a feature in data engineering that automatically loads data from various sources into a data warehouse or database.
Automates the process of loading data from different sources
Reduces manual effort and human error
Can be scheduled to run at specific intervals
Examples: Apache Nifi, AWS Glue
To connect to different services in Azure, you can use Azure SDKs, REST APIs, Azure Portal, Azure CLI, and Azure PowerShell.
Use Azure SDKs for programming languages like Python, Java, C#, etc.
Utilize REST APIs to interact with Azure services programmatically.
Access and manage services through the Azure Portal.
Leverage Azure CLI for command-line interface interactions.
Automate tasks using Azure PowerShell scripts.
Linked Services are connections to external data sources or destinations in Azure Data Factory.
Linked Services define the connection information needed to connect to external data sources or destinations.
They can be used in Data Factory pipelines to read from or write to external systems.
Examples of Linked Services include Azure Blob Storage, Azure SQL Database, and Amazon S3.
I applied via Recruitment Consulltant and was interviewed in Nov 2024. There were 2 interview rounds.
I have a background in data analysis with experience in using tools like Python, SQL, and Tableau.
I have a degree in Statistics and have worked as a Data Analyst for 3 years.
My daily activities include cleaning and analyzing data, creating visualizations, and presenting insights to stakeholders.
I use Python for data manipulation and analysis, SQL for querying databases, and Tableau for creating interactive dashboards.
I...
Advanced Excel and Power BI are tools used for data analysis and visualization in companies and for clients.
Advanced Excel allows for complex data manipulation, analysis, and visualization using features like pivot tables, macros, and VBA programming.
Power BI is a business analytics tool that provides interactive visualizations and business intelligence capabilities, connecting to various data sources.
These tools are u...
I have extensive experience in using Advanced Excel and Power BI for data analysis projects.
Created complex formulas and macros in Excel to automate data processing tasks
Designed interactive dashboards in Power BI to visualize and analyze data trends
Integrated data from multiple sources into Power BI for comprehensive analysis
Used Power Query and Power Pivot in Excel to manipulate and analyze large datasets
Provided dat...
Credit and operations concepts in relation to KYC procedures and client data privacy.
Credit refers to the extension of money or resources to a client based on their financial history and ability to repay.
Operations involve the day-to-day processes and procedures within a financial institution to ensure smooth functioning.
KYC procedures are used to verify the identity of clients to prevent fraud and money laundering.
Pri...
I applied via Job Fair and was interviewed in Nov 2024. There were 2 interview rounds.
I was interviewed in Dec 2024.
I applied via Company Website and was interviewed in Sep 2024. There were 2 interview rounds.
Platform - Hackerank
Duration - 2 Hours
Topics - Spark and SQL
Common file formats used in data storages include CSV, JSON, Parquet, Avro, and ORC. Parquet is best for compression.
CSV (Comma-Separated Values) - simple and widely used, but not efficient for large datasets
JSON (JavaScript Object Notation) - human-readable and easy to parse, but can be inefficient for storage
Parquet - columnar storage format that is highly efficient for compression and query performance
Avro - efficie...
Python program to find the most repeating substring in a list of words.
Iterate through each word in the list
Generate all possible substrings for each word
Count the occurrences of each substring using a dictionary
Find the substring with the highest count
Software Engineer
932
salaries
| ₹2.8 L/yr - ₹11 L/yr |
Senior Associate
762
salaries
| ₹1.2 L/yr - ₹9.3 L/yr |
Network Engineer
647
salaries
| ₹1.8 L/yr - ₹10 L/yr |
Software Developer
615
salaries
| ₹2.5 L/yr - ₹13 L/yr |
Senior Software Engineer
512
salaries
| ₹6.5 L/yr - ₹24 L/yr |
Tata Communications
Bharti Airtel
Reliance Communications
Vodafone Idea