i
LTIMindtree
Filter interviews by
I applied via Approached by Company
Window function coding test involves using window functions in SQL to perform calculations within a specified window of rows.
Understand the syntax and usage of window functions in SQL
Use window functions like ROW_NUMBER(), RANK(), DENSE_RANK(), etc. to perform calculations
Specify the window frame using PARTITION BY and ORDER BY clauses
Practice writing queries with window functions to get comfortable with their usage
Azure Data Factory is a cloud-based data integration service that allows you to create, schedule, and manage data pipelines.
Azure Data Factory is used to move and transform data from various sources to destinations.
It supports data integration processes like ETL (Extract, Transform, Load) and ELT (Extract, Load, Transform).
You can create data pipelines using a visual interface in Azure Data Factory.
It can connect to on...
Data Vault is a modeling methodology for designing highly scalable and flexible data warehouses.
Data Vault focuses on long-term historical data storage
It consists of three main components: Hubs, Links, and Satellites
Hubs represent business entities, Links represent relationships between entities, and Satellites store attributes of entities
Data Vault allows for easy scalability and adaptability to changing business requ
Lambda architecture is a data processing architecture designed to handle massive quantities of data by using both batch and stream processing methods.
Combines batch processing layer, speed layer, and serving layer
Batch layer processes historical data in large batches
Speed layer processes real-time data
Serving layer merges results from batch and speed layers for querying
Example: Apache Hadoop for batch processing, Apach
Yes, I have onsite exposure in previous roles.
I have worked onsite at various client locations to gather requirements and implement solutions.
I have experience collaborating with cross-functional teams in person.
I have conducted onsite training sessions for end users on data architecture best practices.
I have participated in onsite data migration projects.
I have worked onsite to troubleshoot and resolve data-related is
Spark has a master-slave architecture with a cluster manager and worker nodes.
Spark has a driver program that communicates with a cluster manager to allocate resources and schedule tasks.
The cluster manager can be standalone, Mesos, or YARN.
Worker nodes execute tasks and store data in memory or on disk.
Spark can also utilize external data sources like Hadoop Distributed File System (HDFS) or Amazon S3.
Spark supports va...
I applied via Campus Placement and was interviewed before Jan 2021. There were 4 interview rounds.
I have worked on various technologies including Hadoop, Spark, SQL, Python, and AWS.
Experience with Hadoop and Spark for big data processing
Proficient in SQL for data querying and manipulation
Skilled in Python for data analysis and scripting
Familiarity with AWS services such as S3, EC2, and EMR
Knowledge of data warehousing and ETL processes
I applied via Campus Placement and was interviewed before Jul 2021. There were 3 interview rounds.
In this round we have aptitude plus coding mcq questions
Here we have to write full fledge code 2 questions were there and are easy
I applied via Walk-in and was interviewed before Feb 2020. There was 1 interview round.
I applied via Referral and was interviewed before Jun 2021. There were 2 interview rounds.
I applied via Naukri.com and was interviewed in Jan 2024. There was 1 interview round.
I applied via Referral and was interviewed in Jul 2021. There was 1 interview round.
Smb join is a method used to join two tables in SQL Server.
Smb join stands for Sort Merge Bucket join.
It is used when joining large tables.
It involves sorting the tables and then merging them.
It is an efficient join method for large tables with indexes.
Example: SELECT * FROM table1 JOIN table2 ON table1.column = table2.column OPTION (HASH JOIN, MERGE JOIN, LOOP JOIN);
I applied via Naukri.com and was interviewed before Aug 2023. There were 3 interview rounds.
3 SQL question -2 window, join and cte
To connect to ADLS Gen2 from Databricks, you can use the Azure Blob Storage API.
Use the Azure Blob Storage API to connect to ADLS Gen2 from Databricks
Provide the storage account name and key for authentication
Use the storage account name as the filesystem
Example: spark.conf.set('fs.azure.account.key.
Activities in the pipeline include data extraction, transformation, loading, and monitoring.
Data extraction: Retrieving data from various sources such as databases, APIs, and files.
Data transformation: Cleaning, filtering, and structuring the data for analysis.
Data loading: Storing the processed data into a data warehouse or database.
Monitoring: Tracking the pipeline performance, data quality, and handling errors.
ADLS Gen2 offers better performance, security, and scalability compared to Gen1.
ADLS Gen2 supports hierarchical file system, while Gen1 does not
ADLS Gen2 integrates with Azure Blob Storage, providing a unified data lake solution
ADLS Gen2 offers better performance and scalability compared to Gen1
ADLS Gen2 provides better security features such as Azure Data Lake Storage firewall and Azure Active Directory integration
Yes, I have worked on Azure Key Vault (AKV) for securely storing and managing sensitive information.
Implemented AKV to securely store and manage sensitive information such as passwords, certificates, and keys
Utilized AKV to encrypt data at rest and in transit
Integrated AKV with applications for secure access to secrets
Data Engineers are responsible for designing, constructing, installing, and maintaining data management systems.
Designing and implementing data pipelines
Building and maintaining data warehouses
Ensuring data quality and integrity
Collaborating with data scientists and analysts
Optimizing data processes for performance and scalability
Performance tuning in Spark involves optimizing resource allocation and minimizing data shuffling.
Use appropriate cluster configuration and resource allocation
Minimize data shuffling by using appropriate partitioning and caching
Use efficient transformations and actions
Avoid unnecessary operations and transformations
Use broadcast variables for small data sets
Use appropriate serialization formats
Monitor and optimize garb...
based on 1 interview
Interview experience
based on 5 reviews
Rating in categories
Senior Software Engineer
21.3k
salaries
| ₹0 L/yr - ₹0 L/yr |
Software Engineer
16.2k
salaries
| ₹0 L/yr - ₹0 L/yr |
Technical Lead
6.4k
salaries
| ₹0 L/yr - ₹0 L/yr |
Module Lead
5.9k
salaries
| ₹0 L/yr - ₹0 L/yr |
Senior Engineer
4.4k
salaries
| ₹0 L/yr - ₹0 L/yr |
Cognizant
Capgemini
Accenture
TCS