Filter interviews by
I applied via Campus Placement and was interviewed in Jul 2022. There were 4 interview rounds.
SQL coding and aptitude were part of the first round
Top trending discussions
I applied via Recruitment Consulltant and was interviewed in Nov 2024. There were 2 interview rounds.
Various data warehousing techniques include dimensional modeling, star schema, snowflake schema, and data vault.
Dimensional modeling involves organizing data into facts and dimensions to facilitate easy querying and analysis.
Star schema is a type of dimensional modeling where a central fact table is connected to multiple dimension tables.
Snowflake schema is an extension of star schema where dimension tables are normali...
My analytics work has helped the organization make data-driven decisions, improve operational efficiency, and identify new opportunities for growth.
Developed data models and algorithms to optimize business processes
Generated insights from large datasets to drive strategic decision-making
Identified trends and patterns to improve customer experience and retention
Implemented data governance policies to ensure data quality
I would respond in various situations by remaining calm, assessing the situation, and providing a thoughtful and strategic solution.
Remain calm and composed
Assess the situation thoroughly
Provide a thoughtful and strategic solution
Communicate effectively with all parties involved
Both career and team are important, but ultimately career growth should be prioritized.
Career growth is essential for personal development and achieving professional goals.
A strong team can support career growth by providing mentorship, collaboration, and opportunities for learning.
Balancing career and team dynamics is key to long-term success in any role.
I applied via Naukri.com and was interviewed in Jun 2024. There were 3 interview rounds.
I have used HUDI and Iceberg in my previous project for managing large-scale data lakes efficiently.
Implemented HUDI for incremental data ingestion and managing large datasets in real-time
Utilized Iceberg for efficient table management and data versioning
Integrated HUDI and Iceberg with Apache Spark for processing and querying data
I applied via Approached by Company
Window function coding test involves using window functions in SQL to perform calculations within a specified window of rows.
Understand the syntax and usage of window functions in SQL
Use window functions like ROW_NUMBER(), RANK(), DENSE_RANK(), etc. to perform calculations
Specify the window frame using PARTITION BY and ORDER BY clauses
Practice writing queries with window functions to get comfortable with their usage
Azure Data Factory is a cloud-based data integration service that allows you to create, schedule, and manage data pipelines.
Azure Data Factory is used to move and transform data from various sources to destinations.
It supports data integration processes like ETL (Extract, Transform, Load) and ELT (Extract, Load, Transform).
You can create data pipelines using a visual interface in Azure Data Factory.
It can connect to on...
Data Vault is a modeling methodology for designing highly scalable and flexible data warehouses.
Data Vault focuses on long-term historical data storage
It consists of three main components: Hubs, Links, and Satellites
Hubs represent business entities, Links represent relationships between entities, and Satellites store attributes of entities
Data Vault allows for easy scalability and adaptability to changing business requ
Lambda architecture is a data processing architecture designed to handle massive quantities of data by using both batch and stream processing methods.
Combines batch processing layer, speed layer, and serving layer
Batch layer processes historical data in large batches
Speed layer processes real-time data
Serving layer merges results from batch and speed layers for querying
Example: Apache Hadoop for batch processing, Apach
Yes, I have onsite exposure in previous roles.
I have worked onsite at various client locations to gather requirements and implement solutions.
I have experience collaborating with cross-functional teams in person.
I have conducted onsite training sessions for end users on data architecture best practices.
I have participated in onsite data migration projects.
I have worked onsite to troubleshoot and resolve data-related is
I applied via Recruitment Consulltant and was interviewed in May 2024. There were 2 interview rounds.
SQL Scripts to write and also also asked to design an data model of my choice in Telecom Domain
I applied via Naukri.com and was interviewed in Feb 2024. There was 1 interview round.
I was interviewed before Jan 2024.
I applied via Approached by Company and was interviewed before Feb 2023. There was 1 interview round.
A data model for book lending
Create entities for books, borrowers, and loans
Include attributes such as book title, author, borrower name, loan date, and due date
Establish relationships between books and borrowers through loan transactions
Consider additional attributes like book genre, borrower contact information, and loan status
I applied via Company Website and was interviewed before Aug 2021. There was 1 interview round.
Software Engineer
8
salaries
| ₹18.5 L/yr - ₹37 L/yr |
Software Developer
4
salaries
| ₹22 L/yr - ₹25 L/yr |
Devops Engineer
4
salaries
| ₹20 L/yr - ₹22 L/yr |
Software Engineer2
4
salaries
| ₹38 L/yr - ₹93 L/yr |
Software Automation Engineer
3
salaries
| ₹13 L/yr - ₹22.2 L/yr |
CarDekho
Cartrade.com
Carwale
CARS24