Filter interviews by
I applied via Approached by Company and was interviewed before Feb 2023. There were 2 interview rounds.
I applied via Referral and was interviewed in Nov 2022. There were 3 interview rounds.
Top trending discussions
I applied via Recruitment Consulltant and was interviewed in Nov 2024. There were 2 interview rounds.
Different types of joins available in Databricks include inner join, outer join, left join, right join, and cross join.
Inner join: Returns only the rows that have matching values in both tables.
Outer join: Returns all rows when there is a match in either table.
Left join: Returns all rows from the left table and the matched rows from the right table.
Right join: Returns all rows from the right table and the matched rows ...
Implementing fault tolerance in a data pipeline involves redundancy, monitoring, and error handling.
Use redundant components to ensure continuous data flow
Implement monitoring tools to detect failures and bottlenecks
Set up automated alerts for immediate response to issues
Design error handling mechanisms to gracefully handle failures
Use checkpoints and retries to ensure data integrity
AutoLoader is a feature in data engineering that automatically loads data from various sources into a data warehouse or database.
Automates the process of loading data from different sources
Reduces manual effort and human error
Can be scheduled to run at specific intervals
Examples: Apache Nifi, AWS Glue
To connect to different services in Azure, you can use Azure SDKs, REST APIs, Azure Portal, Azure CLI, and Azure PowerShell.
Use Azure SDKs for programming languages like Python, Java, C#, etc.
Utilize REST APIs to interact with Azure services programmatically.
Access and manage services through the Azure Portal.
Leverage Azure CLI for command-line interface interactions.
Automate tasks using Azure PowerShell scripts.
Linked Services are connections to external data sources or destinations in Azure Data Factory.
Linked Services define the connection information needed to connect to external data sources or destinations.
They can be used in Data Factory pipelines to read from or write to external systems.
Examples of Linked Services include Azure Blob Storage, Azure SQL Database, and Amazon S3.
I applied via Recruitment Consulltant and was interviewed in Oct 2024. There was 1 interview round.
Different types of joins in SQL include inner join, left join, right join, and full outer join.
Inner join: Returns rows when there is a match in both tables.
Left join: Returns all rows from the left table and the matched rows from the right table.
Right join: Returns all rows from the right table and the matched rows from the left table.
Full outer join: Returns rows when there is a match in either table.
Various optimizations such as indexing, caching, and parallel processing were used in the project.
Implemented indexing on frequently queried columns to improve query performance
Utilized caching mechanisms to store frequently accessed data and reduce database load
Implemented parallel processing to speed up data processing tasks
Optimized algorithms and data structures for efficient data retrieval and manipulation
I optimized code, increased memory allocation, used efficient data structures, and implemented data partitioning.
Optimized code by identifying and fixing memory leaks
Increased memory allocation for the application
Used efficient data structures like arrays, hashmaps, and trees
Implemented data partitioning to distribute data across multiple nodes
To add a new column to data in Pyspark, use 'withColumn' method. To read data from a CSV file, use 'spark.read.csv' method.
To add a new column to data in Pyspark, use 'withColumn' method
Example: df.withColumn('new_column', df['existing_column'] * 2)
To read data from a CSV file, use 'spark.read.csv' method
Example: df = spark.read.csv('file.csv', header=True, inferSchema=True)
I applied via Amazon jobs and was interviewed in Sep 2023. There were 3 interview rounds.
Basic with python SQL and data models
Developing a data pipeline to analyze customer behavior for an e-commerce company
Collecting and storing customer data from website interactions
Cleaning and transforming data to identify patterns and trends
Building machine learning models to predict customer behavior
Visualizing insights for stakeholders to make data-driven decisions
based on 1 review
Rating in categories
Software Engineer2
18
salaries
| ₹20 L/yr - ₹23 L/yr |
Senior Data Engineer
12
salaries
| ₹23 L/yr - ₹46 L/yr |
Senior Software Engineer
12
salaries
| ₹15.6 L/yr - ₹44.8 L/yr |
Software Engineer
11
salaries
| ₹7.5 L/yr - ₹15 L/yr |
Data Quality Analyst
11
salaries
| ₹7.1 L/yr - ₹9 L/yr |
CBRE
Colliers India
Cushman & Wakefield
Knight Frank