Filter interviews by
I applied via Recruitment Consulltant and was interviewed before Mar 2021. There was 1 interview round.
Java programming questions in hanker rank
I applied via Naukri.com and was interviewed in Jul 2023. There were 2 interview rounds.
Spark internal working and optimization techniques
Spark uses Directed Acyclic Graph (DAG) for optimizing workflows
Lazy evaluation helps in optimizing transformations by combining them into a single stage
Caching and persistence of intermediate results can improve performance
Partitioning data can help in parallel processing and reducing shuffle operations
Spark and Hadoop have different architectures. DAG is a directed acyclic graph. Stage boundaries are logical divisions in a Spark job. Hive has partitioning and bucketing.
Spark is an in-memory processing engine while Hadoop is a distributed file system.
DAG is a graph of stages in a Spark job.
Stage boundaries are logical divisions in a Spark job that help optimize execution.
Partitioning in Hive is a way to divide a tabl...
posted on 16 Nov 2024
I applied via Naukri.com and was interviewed in Nov 2024. There was 1 interview round.
Enhanced optimization in AWS Glue improves job performance by automatically adjusting resources based on workload
Enhanced optimization in AWS Glue automatically adjusts resources like DPUs based on workload
It helps improve job performance by optimizing resource allocation
Users can enable enhanced optimization in AWS Glue job settings
Optimizing querying in Amazon Redshift involves proper table design, distribution keys, sort keys, and query optimization techniques.
Use appropriate distribution keys to evenly distribute data across nodes for parallel processing.
Utilize sort keys to physically order data on disk, reducing the need for sorting during queries.
Avoid using SELECT * and instead specify only the columns needed to reduce data transfer.
Use AN...
I applied via Naukri.com and was interviewed in Sep 2024. There was 1 interview round.
SCD type 2 is a method used in data warehousing to track historical changes by creating a new record for each change.
SCD type 2 stands for Slowly Changing Dimension type 2
It involves creating a new record in the dimension table whenever there is a change in the data
The old record is marked as inactive and the new record is marked as current
It allows for historical tracking of changes in data over time
Example: If a cust...
posted on 9 Jan 2025
I applied via Campus Placement and was interviewed in Jul 2024. There was 1 interview round.
I applied via Job Portal and was interviewed in Jul 2024. There was 1 interview round.
I have over 5 years of experience in data engineering, working with large datasets and implementing data pipelines.
Developed and maintained ETL processes to extract, transform, and load data from various sources
Optimized database performance and implemented data quality checks
Worked with cross-functional teams to design and implement data solutions
Utilized tools such as Apache Spark, Hadoop, and SQL for data processing
...
I would start by understanding the requirements, breaking down the task into smaller steps, researching if needed, and then creating a plan to execute the task efficiently.
Understand the requirements of the task
Break down the task into smaller steps
Research if needed to gather necessary information
Create a plan to execute the task efficiently
Communicate with stakeholders for clarification or updates
Regularly track prog
based on 1 review
Rating in categories
Senior Software Engineer
2.6k
salaries
| ₹15 L/yr - ₹42.7 L/yr |
Software Engineer
1.7k
salaries
| ₹6.9 L/yr - ₹24 L/yr |
Lead Software Engineer
831
salaries
| ₹18 L/yr - ₹52 L/yr |
Senior Systems Engineer
304
salaries
| ₹12 L/yr - ₹36.3 L/yr |
Software Test Automation Engineer
267
salaries
| ₹7 L/yr - ₹20 L/yr |
TCS
Infosys
Wipro
HCLTech