Filter interviews by
Clear (1)
Types of clusters in Databricks include Standard, High Concurrency, and Single Node clusters.
Standard clusters are used for general-purpose workloads
High Concurrency clusters are optimized for concurrent workloads
Single Node clusters are used for development and testing purposes
Catalyst optimizer is a query optimizer in Apache Spark that leverages advanced techniques to optimize and improve the performance of Spark SQL queries.
Catalyst optimizer uses a rule-based and cost-based optimization approach to generate an optimized query plan.
It performs various optimizations such as constant folding, predicate pushdown, and projection pruning to improve query performance.
Catalyst optimizer also leve...
Explode function is used in Apache Spark to split an array into multiple rows.
Used in Apache Spark to split an array into multiple rows
Creates a new row for each element in the array
Commonly used in data processing and transformation tasks
Delta Lake is an open-source storage layer that brings ACID transactions to Apache Spark and big data workloads.
Delta Lake provides ACID transactions, schema enforcement, and time travel capabilities on top of data lakes.
Data lakes are a storage repository that holds a vast amount of raw data in its native format until it is needed.
Delta Lake is optimized for big data workloads and provides reliability and performance ...
RDD stands for Resilient Distributed Dataset, a fundamental data structure in Apache Spark.
RDD is a fault-tolerant collection of elements that can be operated on in parallel.
RDDs are immutable, meaning they cannot be changed once created.
RDDs support two types of operations: transformations (creating a new RDD from an existing one) and actions (returning a value to the driver program).
I applied via Recruitment Consulltant and was interviewed before Apr 2021. There were 3 interview rounds.
Standard Aptitude questions
Based on strings and array
I applied via Referral and was interviewed before Feb 2021. There were 2 interview rounds.
I applied via Campus Placement and was interviewed in Oct 2020. There were 4 interview rounds.
Dictionary is a collection of key-value pairs in Python.
Keys must be unique and immutable.
Values can be of any data type.
Access values using keys.
Add or modify values using keys.
Use dict() constructor or {} to create a dictionary.
Example: {'name': 'John', 'age': 30}
Example: dict(name='John', age=30)
based on 1 interview
Interview experience
Hyderabad / Secunderabad
6-10 Yrs
Not Disclosed
Bangalore / Bengaluru
4-9 Yrs
Not Disclosed
Senior Software Engineer
466
salaries
| ₹0 L/yr - ₹0 L/yr |
Senior Consultant
354
salaries
| ₹0 L/yr - ₹0 L/yr |
Consultant
275
salaries
| ₹0 L/yr - ₹0 L/yr |
Software Engineer
208
salaries
| ₹0 L/yr - ₹0 L/yr |
Senior Software Developer
135
salaries
| ₹0 L/yr - ₹0 L/yr |
Accenture
EPAM Systems
GlobalLogic
LTIMindtree