i
Wipro
Filter interviews by
Clear (1)
ADF questions refer to Azure Data Factory questions which are related to data integration and data transformation processes.
ADF questions are related to Azure Data Factory, a cloud-based data integration service.
These questions may involve data pipelines, data flows, activities, triggers, and data movement.
Candidates may be asked about their experience with designing, monitoring, and managing data pipelines in ADF.
Exam...
Rate your
company
🤫 100% anonymous
How was your last interview experience?
I applied via Naukri.com and was interviewed in Dec 2024. There were 2 interview rounds.
Python coding and SQL questions.
Optimization techniques are methods used to improve the efficiency and performance of data processing.
Use indexing to speed up data retrieval
Implement caching to reduce redundant computations
Utilize parallel processing for faster execution
Optimize algorithms for better performance
Use data partitioning to distribute workload evenly
Merge two unsorted arrays into a single sorted array.
Create a new array to store the merged result
Iterate through both arrays and compare elements to merge in sorted order
Handle remaining elements in either array after one array is fully processed
Apache Spark is a fast and general-purpose cluster computing system.
Apache Spark is an open-source distributed computing system that provides an interface for programming entire clusters with implicit data parallelism and fault tolerance.
It can be used for a wide range of applications such as batch processing, real-time stream processing, machine learning, and graph processing.
Spark provides high-level APIs in Java, Sc...
What people are saying about Wipro
It was more aboiut coding
Wipro interview questions for designations
Get interview-ready with Top Wipro Interview Questions
I applied via Recruitment Consulltant and was interviewed in May 2024. There was 1 interview round.
PySpark is a Python API for Apache Spark, a powerful open-source distributed computing system.
PySpark is used for processing large datasets with distributed computing.
It provides high-level APIs in Python for Spark programming.
PySpark allows seamless integration with Python libraries like Pandas and NumPy.
Example: PySpark can be used for data processing, machine learning, and real-time analytics.
External tables reference data stored outside the database, while internal tables store data within the database.
External tables are defined on data that is stored outside the database, such as in HDFS or S3.
Internal tables store data within the database itself, typically in a managed storage like HDFS or S3.
External tables do not delete data when dropped, while internal tables do.
Internal tables are managed by the dat...
Quant , reasoning, english, coding
I applied via Walk-in
Spark optimization techniques aim to improve performance and efficiency of Spark jobs.
Use partitioning to distribute data evenly
Cache intermediate results to avoid recomputation
Optimize shuffle operations by reducing data shuffling
Use broadcast variables for small lookup tables
Tune memory and executor settings for better performance
based on 25 interviews
2 Interview rounds
Anonymously discuss salaries, work culture, and many more
Get Ambitionbox App
based on 51 reviews
Rating in categories
Chennai,
Bangalore / Bengaluru
5-10 Yrs
₹ 5-12 LPA
Project Engineer
32.8k
salaries
| ₹0 L/yr - ₹0 L/yr |
Senior Software Engineer
23.2k
salaries
| ₹0 L/yr - ₹0 L/yr |
Senior Associate
21.4k
salaries
| ₹0 L/yr - ₹0 L/yr |
Senior Project Engineer
20.3k
salaries
| ₹0 L/yr - ₹0 L/yr |
Technical Lead
18.8k
salaries
| ₹0 L/yr - ₹0 L/yr |
TCS
Infosys
Tesla
Amazon