Filter interviews by
I applied via Company Website and was interviewed in Nov 2024. There was 1 interview round.
Window functions are used to perform calculations across a set of table rows that are related to the current row.
Window functions are used in SQL to perform calculations on a specific subset of rows related to the current row.
They are often used with aggregate functions like SUM, AVG, and COUNT to calculate running totals, moving averages, and rankings.
Examples of window functions include ROW_NUMBER(), RANK(), LEAD(),
Use PySpark code to connect data to ADLS with partitioning
Use SparkSession to create a Spark application
Set the configuration for ADLS storage account and container
Read data from ADLS using Spark DataFrame API
Partition the data based on a specific column while writing back to ADLS
I was interviewed before Oct 2022.
1.ETL Pipeline
2.PySpark Code
3.SQL
I applied via Internshala and was interviewed before Sep 2021. There were 5 interview rounds.
I was given a pdf file containing 3 Problem statements along with the output for which I had to write SQL queries.
Joins are used to combine data from two or more tables based on a related column between them.
Inner Join: returns only the matching rows from both tables
Left Join: returns all rows from the left table and matching rows from the right table
Right Join: returns all rows from the right table and matching rows from the left table
Full Outer Join: returns all rows from both tables
Cross Join: returns the Cartesian product of b
Joins combine rows from two or more tables based on a related column, while unions combine rows from two or more tables with the same structure.
Joins are used to combine data from different tables based on a related column
Unions are used to combine data from tables with the same structure
Joins can be inner, left, right, or full, while unions are always a combination of all rows
Joins can have multiple conditions, while ...
Top trending discussions
Spark has a master-slave architecture with a cluster manager and worker nodes.
Spark has a driver program that communicates with a cluster manager to allocate resources and schedule tasks.
The cluster manager can be standalone, Mesos, or YARN.
Worker nodes execute tasks and store data in memory or on disk.
Spark can also utilize external data sources like Hadoop Distributed File System (HDFS) or Amazon S3.
Spark supports va...
I applied via Walk-in and was interviewed before Feb 2020. There was 1 interview round.
I applied via Campus Placement and was interviewed before Jul 2020. There was 1 interview round.
I applied via Naukri.com and was interviewed in Jan 2024. There was 1 interview round.
I applied via Approached by Company
Implement a Word Count program in Spark Scala
Use Spark's RDD API to read input text file
Split each line into words and map them to key-value pairs
ReduceByKey operation to count occurrences of each word
Save the result to an output file
Higher Order Functions in Scala are functions that take other functions as parameters or return functions as results.
Higher Order Functions allow for more concise and readable code.
Examples include map, filter, reduce, and flatMap in Scala.
They promote code reusability and modularity.
Higher Order Functions are a key feature of functional programming.
I applied via Referral and was interviewed before Oct 2022. There were 4 interview rounds.
I applied via Approached by Company and was interviewed in Jun 2024. There were 2 interview rounds.
based on 1 interview
Interview experience
based on 113 reviews
Rating in categories
5-9 Yrs
Not Disclosed
Data Engineer
56
salaries
| ₹0 L/yr - ₹0 L/yr |
Software Engineer
29
salaries
| ₹0 L/yr - ₹0 L/yr |
Business Analyst
24
salaries
| ₹0 L/yr - ₹0 L/yr |
Senior Software Engineer
18
salaries
| ₹0 L/yr - ₹0 L/yr |
Data Analyst
12
salaries
| ₹0 L/yr - ₹0 L/yr |
TCS
Infosys
Wipro
HCLTech