i
LTIMindtree
Filter interviews by
Clear (1)
I applied via Approached by Company
Window function coding test involves using window functions in SQL to perform calculations within a specified window of rows.
Understand the syntax and usage of window functions in SQL
Use window functions like ROW_NUMBER(), RANK(), DENSE_RANK(), etc. to perform calculations
Specify the window frame using PARTITION BY and ORDER BY clauses
Practice writing queries with window functions to get comfortable with their usage
Azure Data Factory is a cloud-based data integration service that allows you to create, schedule, and manage data pipelines.
Azure Data Factory is used to move and transform data from various sources to destinations.
It supports data integration processes like ETL (Extract, Transform, Load) and ELT (Extract, Load, Transform).
You can create data pipelines using a visual interface in Azure Data Factory.
It can connect to on...
Data Vault is a modeling methodology for designing highly scalable and flexible data warehouses.
Data Vault focuses on long-term historical data storage
It consists of three main components: Hubs, Links, and Satellites
Hubs represent business entities, Links represent relationships between entities, and Satellites store attributes of entities
Data Vault allows for easy scalability and adaptability to changing business requ
Lambda architecture is a data processing architecture designed to handle massive quantities of data by using both batch and stream processing methods.
Combines batch processing layer, speed layer, and serving layer
Batch layer processes historical data in large batches
Speed layer processes real-time data
Serving layer merges results from batch and speed layers for querying
Example: Apache Hadoop for batch processing, Apach
Yes, I have onsite exposure in previous roles.
I have worked onsite at various client locations to gather requirements and implement solutions.
I have experience collaborating with cross-functional teams in person.
I have conducted onsite training sessions for end users on data architecture best practices.
I have participated in onsite data migration projects.
I have worked onsite to troubleshoot and resolve data-related is
I applied via Approached by Company and was interviewed in Dec 2023. There was 1 interview round.
The 7 layers in Azure Data Factory for pipelining data from on-premises to Azure Cloud
1. Ingestion Layer: Collects data from various sources such as on-premises databases, cloud storage, or IoT devices.
2. Storage Layer: Stores the ingested data in a data lake or data warehouse for processing.
3. Batch Layer: Processes data in batches using technologies like Azure Databricks or HDInsight.
4. Stream Layer: Processes real-t...
To convert a normal file to a flat file in Python, you can read the file line by line and write the data to a new file with a delimiter.
Open the normal file in read mode
Read the file line by line
Split the data based on the delimiter (if applicable)
Write the data to a new file with a delimiter
I applied via Campus Placement and was interviewed before Jul 2021. There were 3 interview rounds.
In this round we have aptitude plus coding mcq questions
Here we have to write full fledge code 2 questions were there and are easy
I applied via Campus Placement and was interviewed before Jan 2021. There were 4 interview rounds.
I have worked on various technologies including Hadoop, Spark, SQL, Python, and AWS.
Experience with Hadoop and Spark for big data processing
Proficient in SQL for data querying and manipulation
Skilled in Python for data analysis and scripting
Familiarity with AWS services such as S3, EC2, and EMR
Knowledge of data warehousing and ETL processes
I applied via Walk-in and was interviewed before Feb 2020. There was 1 interview round.
I applied via Referral and was interviewed before Jun 2021. There were 2 interview rounds.
Spark has a master-slave architecture with a cluster manager and worker nodes.
Spark has a driver program that communicates with a cluster manager to allocate resources and schedule tasks.
The cluster manager can be standalone, Mesos, or YARN.
Worker nodes execute tasks and store data in memory or on disk.
Spark can also utilize external data sources like Hadoop Distributed File System (HDFS) or Amazon S3.
Spark supports va...
I applied via Naukri.com
I have worked on various AWS services including S3, EC2, Lambda, Glue, and Redshift.
S3 - Used for storing and retrieving data
EC2 - Used for running virtual servers
Lambda - Used for serverless computing
Glue - Used for ETL (Extract, Transform, Load) processes
Redshift - Used for data warehousing and analytics
I applied via Naukri.com and was interviewed in Jan 2024. There was 1 interview round.
based on 2 interviews
Interview experience
based on 5 reviews
Rating in categories
Senior Software Engineer
21.3k
salaries
| ₹5.1 L/yr - ₹18.8 L/yr |
Software Engineer
16.2k
salaries
| ₹2 L/yr - ₹10 L/yr |
Technical Lead
6.4k
salaries
| ₹9.4 L/yr - ₹36.6 L/yr |
Module Lead
5.9k
salaries
| ₹7 L/yr - ₹25.5 L/yr |
Senior Engineer
4.4k
salaries
| ₹4.2 L/yr - ₹16.5 L/yr |
Cognizant
Capgemini
Accenture
TCS