i
Sigmoid
Filter interviews by
Find indices of an element in a non-decreasing array
Iterate through the array and keep track of the indices where the element X is found
Return the list of indices or [-1, -1] if element X is not found
Handle edge cases like empty array or X not present in the array
I applied via Naukri.com and was interviewed in Sep 2024. There was 1 interview round.
SCD stands for Slowly Changing Dimension in Data Warehousing.
SCD is a technique used in data warehousing to track changes to dimension data over time.
There are different types of SCDs - Type 1, Type 2, and Type 3.
Type 1 SCD overwrites old data with new data, Type 2 creates new records for changes, and Type 3 maintains both old and new values in separate columns.
Example: In a customer dimension table, if a customer chan...
inferschema in pyspark is used to automatically infer the schema of a file when reading it.
inferschema is a parameter in pyspark that can be set to true when reading a file to automatically infer the schema based on the data
It is useful when the schema of the file is not known beforehand
Example: df = spark.read.csv('file.csv', header=True, inferSchema=True)
Rank assigns unique ranks to each distinct value, while dense rank assigns ranks without gaps.
Rank function assigns unique ranks to each distinct value in a result set.
Dense rank function assigns ranks to rows in a result set without any gaps between the ranks.
Rank function may skip ranks if there are ties in values, while dense rank will not skip ranks.
Optimizing techniques in Spark involve partitioning, caching, and tuning resources for efficient data processing.
Use partitioning to distribute data evenly across nodes for parallel processing
Cache frequently accessed data in memory to avoid recomputation
Tune resources such as memory allocation and parallelism settings for optimal performance
Repartition is used to increase the number of partitions in a DataFrame, while coalesce is used to decrease the number of partitions.
Repartition involves shuffling data across the network, which can be expensive in terms of performance and resources.
Coalesce is a more efficient operation as it minimizes data movement by only merging existing partitions.
Repartition is typically used when there is a need for more paralle...
Normalization in databases is the process of organizing data in a database to reduce redundancy and improve data integrity.
Normalization is used to eliminate redundant data and ensure data integrity.
It involves breaking down a table into smaller tables and defining relationships between them.
There are different normal forms such as 1NF, 2NF, 3NF, and BCNF.
Normalization helps in reducing data redundancy and improving qu...
Transformation involves changing the data structure, while action involves performing a computation on the data.
Transformation changes the data structure without executing any computation
Action performs a computation on the data and triggers the execution
Examples of transformation include map, filter, and reduce in Spark or Pandas
Examples of action include count, collect, and saveAsTextFile in Spark
Share price related question.most profit
Top trending discussions
I applied via Approached by Company and was interviewed before Jul 2021. There was 1 interview round.
To write a REST API from scratch, I would follow these steps:
Define the resources and endpoints
Choose a programming language and framework
Implement CRUD operations for each resource
Use HTTP methods and status codes correctly
Add authentication and authorization
Test the API using tools like Postman
Document the API using tools like Swagger
posted on 3 Nov 2022
I applied via Campus Placement and was interviewed before Nov 2021. There were 3 interview rounds.
There are 30 basic aptitude question in 30 minutes
3 coding question 2 were easy and 1 was of medium level
First round was online coding round, second was coding in interview infornt of panel, third was DSA+basic Database questions
I appeared for an interview in Apr 2024.
I applied via LinkedIn and was interviewed in Oct 2021. There were 4 interview rounds.
I applied via Recruitment Consulltant and was interviewed in Feb 2024. There was 1 interview round.
Basic ds algo round with questions asked around arrays strings
based on 3 interviews
Interview experience
based on 2 reviews
Rating in categories
Software Development Engineer II
91
salaries
| ₹15 L/yr - ₹23 L/yr |
Data Engineer
63
salaries
| ₹6.5 L/yr - ₹25 L/yr |
Data Scientist
58
salaries
| ₹7.2 L/yr - ₹24 L/yr |
Software Development Engineer
42
salaries
| ₹13.2 L/yr - ₹19.4 L/yr |
Senior Data Scientist
41
salaries
| ₹15 L/yr - ₹28.9 L/yr |
Crisil
Indegene
Kantar
Acuity Knowledge Partners