Upload Button Icon Add office photos

Filter interviews by

Barclays Shared Services Data Engineer Interview Questions and Answers

Updated 19 Jul 2024

Barclays Shared Services Data Engineer Interview Experiences

1 interview found

Data Engineer Interview Questions & Answers

user image Anonymous

posted on 19 Jul 2024

Interview experience
5
Excellent
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Technical 

(2 Questions)

  • Q1. Explain Hadoop architecture?
  • Ans. 

    Hadoop architecture is a distributed computing framework for processing large data sets across clusters of computers.

    • Hadoop consists of HDFS (Hadoop Distributed File System) for storage and MapReduce for processing.

    • HDFS divides data into blocks and stores them across multiple nodes in a cluster.

    • MapReduce is a programming model for processing large data sets in parallel across a distributed cluster.

    • Hadoop also includes ...

  • Answered by AI
  • Q2. Hadoop vs spark difference
  • Ans. 

    Hadoop is a distributed storage system while Spark is a distributed processing engine.

    • Hadoop is primarily used for storing and processing large volumes of data in a distributed environment.

    • Spark is designed for fast data processing and can perform in-memory computations, making it faster than Hadoop for certain tasks.

    • Hadoop uses MapReduce for processing data, while Spark uses Resilient Distributed Datasets (RDDs) for f...

  • Answered by AI

Skills evaluated in this interview

Data Engineer Jobs at Barclays Shared Services

View all

Interview questions from similar companies

Interview experience
3
Average
Difficulty level
-
Process Duration
-
Result
No response
Round 1 - Technical 

(4 Questions)

  • Q1. What is the architecture of Apache Spark?
  • Ans. 

    Apache Spark architecture includes a cluster manager, worker nodes, and driver program.

    • Apache Spark architecture consists of a cluster manager, which allocates resources and schedules tasks.

    • Worker nodes execute tasks and store data in memory or disk.

    • Driver program coordinates tasks and communicates with the cluster manager.

    • Spark applications run as independent sets of processes on a cluster, coordinated by the SparkCon...

  • Answered by AI
  • Q2. What is the difference between the reduceBy and groupBy transformations in Apache Spark?
  • Ans. 

    reduceBy is used to aggregate data based on key, while groupBy is used to group data based on key.

    • reduceBy is a transformation that combines the values of each key using an associative function and a neutral 'zero value'.

    • groupBy is a transformation that groups the data based on a key and returns a grouped data set.

    • reduceBy is more efficient for aggregating data as it reduces the data before shuffling, while groupBy shu...

  • Answered by AI
  • Q3. What is the difference between RDD (Resilient Distributed Datasets) and DataFrame in Apache Spark?
  • Ans. 

    RDD is a low-level abstraction representing a distributed collection of objects, while DataFrame is a higher-level abstraction representing a distributed collection of data organized into named columns.

    • RDD is more suitable for unstructured data and low-level transformations, while DataFrame is more suitable for structured data and high-level abstractions.

    • DataFrames provide optimizations like query optimization and code...

  • Answered by AI
  • Q4. What are the different modes of execution in Apache Spark?
  • Ans. 

    The different modes of execution in Apache Spark include local mode, standalone mode, YARN mode, and Mesos mode.

    • Local mode: Spark runs on a single machine with one executor.

    • Standalone mode: Spark runs on a cluster managed by a standalone cluster manager.

    • YARN mode: Spark runs on a Hadoop cluster using YARN as the resource manager.

    • Mesos mode: Spark runs on a Mesos cluster with Mesos as the resource manager.

  • Answered by AI
Interview experience
3
Average
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
-

I applied via Approached by Company and was interviewed in Apr 2024. There was 1 interview round.

Round 1 - Technical 

(8 Questions)

  • Q1. Python Coding Question 1: Make a list with all datatypes objects in it.
  • Ans. 

    Create a list containing all Python data types.

    • Use the following data types: int, float, complex, str, list, tuple, dict, set, bool, bytes, bytearray, memoryview, None

    • Example: ['int', 'float', 'complex', 'str', 'list', 'tuple', 'dict', 'set', 'bool', 'bytes', 'bytearray', 'memoryview', 'None']

  • Answered by AI
  • Q2. Python Coding Question 2: Extract a character from the string present in the list.
  • Ans. 

    Extract a character from a string in a list of strings.

    • Iterate through the list of strings

    • Use indexing to extract the desired character from each string

    • Handle cases where the index is out of range

    • Return the extracted characters as a new list

  • Answered by AI
  • Q3. Python Coding Question 3: Make a dictionary with Name and Age having 4 records.
  • Ans. 

    Create a dictionary with Name and Age for 4 records in Python.

    • Use curly braces {} to create a dictionary.

    • Separate key-value pairs with a colon :

    • Separate each record with a comma ,

  • Answered by AI
  • Q4. Python Coding Question 4: Make a string and write code to return the output with the reverse string.
  • Q5. Python Coding Question 5: Write a function to check if the string is Palindrome. Check the function with a Palindrome.
  • Ans. 

    Function to check if a string is a palindrome.

    • Create a function that takes a string as input.

    • Reverse the string and compare it with the original string.

    • Return true if they are the same, false otherwise.

    • Example: 'racecar' is a palindrome.

  • Answered by AI
  • Q6. SQL Coding Question 1: Write an SQL query to extract the data from employees table, extract day of joining, month of joining, year of joining from the "Date Of Joining".
  • Q7. What is Data normalization and standardization?
  • Ans. 

    Data normalization is the process of organizing data in a database efficiently, while data standardization is the process of ensuring consistency and uniformity in data.

    • Data normalization involves organizing data into tables and columns to reduce redundancy and improve data integrity.

    • Data standardization involves ensuring that data is consistent and uniform across the database.

    • Normalization helps in reducing data redun...

  • Answered by AI
  • Q8. Statistics/Probability Question: 4 ball of each colors(Red, Green, Blue) are present in the box. After drawing 3 ball randomly from the box, what is the probability of getting all 3 ball having the same co...
  • Ans. 

    The probability of drawing 3 balls of the same color from a box containing 4 balls of each color (Red, Green, Blue).

    • Calculate the total number of ways to draw 3 balls out of 12 balls

    • Calculate the number of ways to draw 3 balls of the same color

    • Divide the number of favorable outcomes by the total number of outcomes to get the probability

  • Answered by AI
Interview experience
5
Excellent
Difficulty level
-
Process Duration
-
Result
-
Round 1 - One-on-one 

(2 Questions)

  • Q1. What is spark and its architecture
  • Ans. 

    Apache Spark is a fast and general-purpose cluster computing system.

    • Apache Spark is an open-source distributed computing system that provides an interface for programming entire clusters with implicit data parallelism and fault tolerance.

    • It has a unified architecture that combines SQL, streaming, machine learning, and graph processing capabilities.

    • Spark architecture consists of a driver program that coordinates the exe...

  • Answered by AI
  • Q2. Small question on ADF
Round 2 - One-on-one 

(2 Questions)

  • Q1. Scenerio based question on Databricks
  • Q2. Scenerio based on ADF

Skills evaluated in this interview

Interview experience
4
Good
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Technical 

(1 Question)

  • Q1. Python SQL question
Round 2 - Technical 

(1 Question)

  • Q1. More on Project side
Round 3 - HR 

(1 Question)

  • Q1. Salary Discussion
Interview experience
1
Bad
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Technical 

(1 Question)

  • Q1. Data engineer roles and resposibilities

Interview Preparation Tips

Interview preparation tips for other job seekers - Don't Go, worst management service and lots of office politics.
Interview experience
2
Poor
Difficulty level
-
Process Duration
Less than 2 weeks
Result
Not Selected

I applied via Approached by Company and was interviewed in Apr 2024. There was 1 interview round.

Round 1 - One-on-one 

(3 Questions)

  • Q1. What Volume of data have you handled in your POCs ?
  • Ans. 

    I have handled terabytes of data in my POCs, including data from various sources and formats.

    • Handled terabytes of data in POCs

    • Worked with data from various sources and formats

    • Used tools like Hadoop, Spark, and SQL for data processing

  • Answered by AI
  • Q2. When will you decide to use repartition and coalesce?
  • Ans. 

    Repartition is used for increasing partitions for parallelism, while coalesce is used for decreasing partitions to reduce shuffling.

    • Repartition is used when there is a need for more partitions to increase parallelism.

    • Coalesce is used when there are too many partitions and need to reduce them to avoid shuffling.

    • Example: Repartition can be used before a join operation to evenly distribute data across partitions for bette...

  • Answered by AI
  • Q3. How will you design/configure a cluster if you have given 10 petabytes of data.
  • Ans. 

    Designing/configuring a cluster for 10 petabytes of data involves considerations for storage capacity, processing power, network bandwidth, and fault tolerance.

    • Consider using a distributed file system like HDFS or object storage like Amazon S3 to store and manage the large volume of data.

    • Implement a scalable processing framework like Apache Spark or Hadoop to efficiently process and analyze the data in parallel.

    • Utilize...

  • Answered by AI

Interview Preparation Tips

Interview preparation tips for other job seekers - Sometimes you can easily detect that they want to rush your interview by asking irrelevant questions even though you have not mentioned such expertise in your resume. Also they have only shortlisted the resume just to reject. This breaks the confidence of the person who is looking for the job and honestly conveying the level of expertise he/she has.

Skills evaluated in this interview

Interview experience
5
Excellent
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Selected Selected

I applied via Job Portal and was interviewed in Aug 2024. There were 3 interview rounds.

Round 1 - Aptitude Test 

Its mandatory test even for experience people

Round 2 - Technical 

(1 Question)

  • Q1. Related to technology
Round 3 - HR 

(1 Question)

  • Q1. Very good discussion towards work culture, salary and all
Interview experience
4
Good
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Technical 

(2 Questions)

  • Q1. Sql questions were
  • Q2. Spark related aswell
Interview experience
4
Good
Difficulty level
Moderate
Process Duration
-
Result
Selected Selected

I applied via campus placement at KLS Institute of Management Education and Research, Belgaum and was interviewed in Jun 2024. There were 2 interview rounds.

Round 1 - Coding Test 

Basic coding test like prime

Round 2 - HR 

(2 Questions)

  • Q1. Your Strengths and weakness
  • Q2. Why Zensar why not other

Barclays Shared Services Interview FAQs

How many rounds are there in Barclays Shared Services Data Engineer interview?
Barclays Shared Services interview process usually has 1 rounds. The most common rounds in the Barclays Shared Services interview process are Technical.
How to prepare for Barclays Shared Services Data Engineer interview?
Go through your CV in detail and study all the technologies mentioned in your CV. Prepare at least two technologies or languages in depth if you are appearing for a technical interview at Barclays Shared Services. The most common topics and skills that interviewers at Barclays Shared Services expect are Operations, Service Excellence, Machine Learning, Python and Analysts.
What are the top questions asked in Barclays Shared Services Data Engineer interview?

Some of the top questions asked at the Barclays Shared Services Data Engineer interview -

  1. Explain Hadoop architectu...read more
  2. Hadoop vs spark differe...read more

Tell us how to improve this page.

Data Engineer

Pune

1-4 Yrs

Not Disclosed

Data Engineer - AWS/PySpark

Pune,

Chennai

1-2 Yrs

Not Disclosed

Data Engineer

Pune

1-9 Yrs

Not Disclosed

Explore more jobs
Process Advisor
705 salaries
unlock blur

₹1 L/yr - ₹7 L/yr

Assistant Manager
504 salaries
unlock blur

₹7.5 L/yr - ₹22 L/yr

Assistant Vice President
435 salaries
unlock blur

₹13.5 L/yr - ₹39 L/yr

Senior Analyst
343 salaries
unlock blur

₹3.5 L/yr - ₹10.1 L/yr

Process Associate
197 salaries
unlock blur

₹1 L/yr - ₹6.6 L/yr

Explore more salaries
Compare Barclays Shared Services with

HSBC Group

4.0
Compare

JPMorgan Chase & Co.

4.0
Compare

Standard Chartered

3.8
Compare

Deutsche Bank

3.9
Compare

Calculate your in-hand salary

Confused about how your in-hand salary is calculated? Enter your annual salary (CTC) and get your in-hand salary
Did you find this page helpful?
Yes No
write
Share an Interview