Filter interviews by
Apache Spark architecture includes a cluster manager, worker nodes, and driver program.
Apache Spark architecture consists of a cluster manager, which allocates resources and schedules tasks.
Worker nodes execute tasks and store data in memory or disk.
Driver program coordinates tasks and communicates with the cluster manager.
Spark applications run as independent sets of processes on a cluster, coordinated by the SparkCon...
reduceBy is used to aggregate data based on key, while groupBy is used to group data based on key.
reduceBy is a transformation that combines the values of each key using an associative function and a neutral 'zero value'.
groupBy is a transformation that groups the data based on a key and returns a grouped data set.
reduceBy is more efficient for aggregating data as it reduces the data before shuffling, while groupBy shu...
RDD is a low-level abstraction representing a distributed collection of objects, while DataFrame is a higher-level abstraction representing a distributed collection of data organized into named columns.
RDD is more suitable for unstructured data and low-level transformations, while DataFrame is more suitable for structured data and high-level abstractions.
DataFrames provide optimizations like query optimization and code...
The different modes of execution in Apache Spark include local mode, standalone mode, YARN mode, and Mesos mode.
Local mode: Spark runs on a single machine with one executor.
Standalone mode: Spark runs on a cluster managed by a standalone cluster manager.
YARN mode: Spark runs on a Hadoop cluster using YARN as the resource manager.
Mesos mode: Spark runs on a Mesos cluster with Mesos as the resource manager.
PySpark is a Python API for Apache Spark, used for big data processing and analytics.
PySpark is a Python API for Apache Spark, a fast and general-purpose cluster computing system.
It allows for easy integration with Python libraries and provides high-level APIs in Python.
PySpark can be used for processing large datasets, machine learning, real-time data streaming, and more.
It supports various data sources such as HDFS, ...
PySpark is a Python API for Apache Spark, while Python is a general-purpose programming language.
PySpark is specifically designed for big data processing using Spark, while Python is a versatile programming language used for various applications.
PySpark allows for distributed computing and parallel processing, while Python is primarily used for sequential programming.
PySpark provides libraries and tools for working wit...
WITH clause in SQL is used to create temporary named result sets that can be referenced within the main query.
WITH clause is used to improve the readability and maintainability of complex SQL queries.
It allows creating subqueries or common table expressions (CTEs) that can be referenced multiple times.
The result sets created using WITH clause can be used for recursive queries, data transformation, or simplifying comple...
What people are saying about Virtusa Consulting Services
I applied via Campus Placement and was interviewed before Jan 2022. There were 2 interview rounds.
Coding Test and Web Development
I have worked on various projects involving data engineering, including building data pipelines and optimizing data storage.
Built a data pipeline using Apache Kafka and Apache Spark to process and analyze real-time streaming data.
Optimized data storage by implementing data partitioning and indexing techniques in a large-scale data warehouse.
Developed ETL processes to extract data from various sources, transform it, and...
I am a data engineer with experience in designing and implementing data pipelines for large-scale projects.
Experienced in building and optimizing data pipelines using tools like Apache Spark and Hadoop
Proficient in programming languages like Python and SQL
Skilled in data modeling and database design
Familiar with cloud platforms like AWS and GCP
Strong problem-solving and analytical skills
Effective communicator and team
Virtusa Consulting Services interview questions for designations
posted on 16 Sep 2021
I appeared for an interview before Sep 2020.
Round duration - 60 Minutes
Round difficulty - Medium
Timing was in the morning around 10 am. This round was conducted on Hackerrank and contained 2 coding questions of easy to medium level which varied for all the students taking the test. There were 20 MCQs from topics like Java, Operating Systems, Data structures, and SQL.
I also gave a test for this company last year during internships. At that time MCQs were based on aptitude and reasoning. Also at that time, there were 2 coding questions and timing was dividing according to sections.
Given two integers X
and Y
as the first two numbers of a series, and an integer N
, determine the Nth element of the series following the Fibonacci rule: f(x) = f(x...
Calculate the Nth element of a modified Fibonacci series given the first two numbers and N, with the result modulo 10^9 + 7.
Implement a function to calculate the Nth element of the series using the Fibonacci rule f(x) = f(x - 1) + f(x - 2)
Return the answer modulo 10^9 + 7 due to the possibility of a very large result
The series starts with the first two numbers X and Y, and the position N in the series
Given a string STR
comprising uppercase and lowercase characters and spaces, your task is to count the number of consonants in the string.
A consonant is defined as a...
Count the number of consonants in a given string containing uppercase and lowercase characters and spaces.
Iterate through each character in the string and check if it is a consonant (not a vowel).
Keep a count of the consonants encountered while iterating through the string.
Return the total count of consonants at the end.
Round duration - 45 Minutes
Round difficulty - Medium
This round was conducted on Cisco WebEx and the interviewer was very friendly. Firstly, we were asked briefly to introduce ourselves. Then he moved on to the CV and asked various questions about the projects, certifications, and internships mentioned. He asked about the implementation of a few features in some of the projects and also about the challenges I faced while doing the project and how I finally overcome the challenge. He also asked about various skills mentioned in the resume.
You are provided with two sorted linked lists. Your task is to merge them into a single sorted linked list and return the head of the combined linked list.
...Merge two sorted linked lists into a single sorted linked list with constant space complexity and linear time complexity.
Create a dummy node to start the merged list
Compare the values of the two linked lists and append the smaller value to the merged list
Move the pointer of the merged list and the pointer of the smaller value's linked list
Continue this process until one of the linked lists is fully traversed
Append the ...
Round duration - 30 Minutes
Round difficulty - Easy
This round started with my Introduction. Then he moved on to behavioral questions just to get an understanding of how we deal with situations and our thinking process.
Tip 1 : Do at least 2 good projects and make an attractive resume mentioning all your skills in the best possible way(Don't include any False commitments)
Tip 2 : Practice easy to medium level coding questions from geeks for geeks or Leetcode.
Tip 3 : Prepare everything mentioned in your resume properly including Java, SQL.
Tip 1 : Include projects and experiences in detail.
Tip 2 : Make an eye-catching resume clearly mentioning all your skills including subjects like data structures, JAVA, SQL, etc. The database is a very important skill if you have its knowledge.
I applied via Campus Placement and was interviewed in May 2020. There were 6 interview rounds.
Design patterns are reusable solutions to common software problems. Android system uses various design patterns.
MVC (Model-View-Controller) pattern is used in Android to separate UI logic from business logic.
Singleton pattern is used to ensure only one instance of a class is created.
Observer pattern is used to notify changes in data to multiple components.
Builder pattern is used to simplify complex object creation.
Adap...
Remove duplicate objects from an array of multiple objects.
Create a new array to store unique objects
Loop through the original array and check if the object already exists in the new array
If not, add it to the new array
Return the new array
Maintain state by using a global variable or a state management system.
Use a global variable to store the state and access it whenever needed.
Use a state management system like Redux or MobX to manage the state.
Avoid using local variables or closures to store state as they will be lost when the code is repeated.
Consider using object-oriented programming principles to encapsulate state within objects.
Use functional prog...
posted on 2 Aug 2022
I applied via Naukri.com and was interviewed before Aug 2021. There were 2 interview rounds.
Some of the top questions asked at the Virtusa Consulting Services Data Engineer interview -
based on 4 interviews
Interview experience
based on 21 reviews
Rating in categories
Senior Consultant
3.8k
salaries
| ₹8 L/yr - ₹30 L/yr |
Lead Consultant
3.5k
salaries
| ₹10.5 L/yr - ₹34.6 L/yr |
Software Engineer
3.3k
salaries
| ₹3.5 L/yr - ₹13.3 L/yr |
Consultant
3.3k
salaries
| ₹6 L/yr - ₹20 L/yr |
Associate Consultant
2.7k
salaries
| ₹4.7 L/yr - ₹15.1 L/yr |
Cognizant
TCS
Infosys
Accenture