i
NTT Data
Filter interviews by
Top trending discussions
The aptitude test lasts 30 minutes and focuses on topics relevant to data engineering, including Spark, SQL, Azure, and PySpark.
The coding test is a one-hour examination on PySpark.
I applied via Naukri.com and was interviewed in Nov 2024. There were 2 interview rounds.
I was interviewed in Dec 2024.
I was interviewed in Aug 2024.
Python and sql tasks
I was interviewed in Sep 2024.
Pyspark is a Python library for big data processing using Spark framework.
Pyspark is used for processing large datasets in parallel.
It provides APIs for data manipulation, querying, and analysis.
Example: Using pyspark to read a CSV file and perform data transformations.
Databricks optimisation techniques improve performance and efficiency of data processing on the Databricks platform.
Use cluster sizing and autoscaling to optimize resource allocation based on workload
Leverage Databricks Delta for optimized data storage and processing
Utilize caching and persisting data to reduce computation time
Optimize queries by using appropriate indexing and partitioning strategies
Databricks is a unified data analytics platform that provides a collaborative environment for data engineers.
Databricks is built on top of Apache Spark and provides a workspace for data engineering tasks.
It allows for easy integration with various data sources and tools for data processing.
Databricks provides features like notebooks, clusters, and libraries for efficient data engineering workflows.
posted on 23 Dec 2024
I applied via Naukri.com and was interviewed in Jun 2024. There were 3 interview rounds.
Sample data and its transformations
Sample data can be in the form of CSV, JSON, or database tables
Transformations include cleaning, filtering, aggregating, and joining data
Examples: converting date formats, removing duplicates, calculating averages
Seeking new challenges and opportunities for growth in a more dynamic environment.
Looking for new challenges and opportunities for growth
Seeking a more dynamic work environment
Interested in expanding skill set and knowledge
Want to work on more innovative projects
I applied via Naukri.com and was interviewed in Sep 2024. There was 1 interview round.
SCD type 2 is a method used in data warehousing to track historical changes by creating a new record for each change.
SCD type 2 stands for Slowly Changing Dimension type 2
It involves creating a new record in the dimension table whenever there is a change in the data
The old record is marked as inactive and the new record is marked as current
It allows for historical tracking of changes in data over time
Example: If a cust...
posted on 4 Aug 2024
I am a Senior Data Engineer with 5+ years of experience in designing and implementing data pipelines for large-scale projects.
Experienced in ETL processes and data warehousing
Proficient in programming languages like Python, SQL, and Java
Skilled in working with big data technologies such as Hadoop, Spark, and Kafka
Strong understanding of data modeling and database management
Excellent problem-solving and communication sk
Developing a real-time data processing system for analyzing customer behavior on e-commerce platform.
Utilizing Apache Kafka for real-time data streaming
Implementing Spark for data processing and analysis
Creating machine learning models for customer segmentation
Integrating with Elasticsearch for data indexing and search functionality
posted on 26 Oct 2024
I applied via Naukri.com and was interviewed in Sep 2024. There was 1 interview round.
Spark Optimization, Transformation, DLT, DL, Data Governance
Python
SQL
based on 1 interview
Interview experience
based on 1 review
Rating in categories
Software Engineer
935
salaries
| ₹2.8 L/yr - ₹11 L/yr |
Senior Associate
774
salaries
| ₹1.2 L/yr - ₹7.5 L/yr |
Network Engineer
672
salaries
| ₹1.8 L/yr - ₹10 L/yr |
Software Developer
641
salaries
| ₹3.1 L/yr - ₹12 L/yr |
Senior Software Engineer
574
salaries
| ₹6.5 L/yr - ₹26 L/yr |
Tata Communications
Bharti Airtel
Reliance Communications
Vodafone Idea