Upload Button Icon Add office photos
Engaged Employer

i

This company page is being actively managed by Optum Global Solutions Team. If you also belong to the team, you can get access from here

Optum Global Solutions Verified Tick

Compare button icon Compare button icon Compare
4.0

based on 5.8k Reviews

Proud winner of ABECA 2024 - AmbitionBox Employee Choice Awards

zig zag pattern zig zag pattern

Filter interviews by

Optum Global Solutions Data Engineering Manager Interview Questions and Answers

Updated 15 Mar 2024

Optum Global Solutions Data Engineering Manager Interview Experiences

1 interview found

Interview experience
5
Excellent
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Selected Selected

I applied via Referral and was interviewed before Mar 2023. There were 2 interview rounds.

Round 1 - Technical 

(1 Question)

  • Q1. Questions related to SAS
Round 2 - HR 

(2 Questions)

  • Q1. Salary Expectations
  • Q2. Why do you want to change

Data Engineering Manager Jobs at Optum Global Solutions

View all

Interview questions from similar companies

Interview experience
4
Good
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Technical 

(5 Questions)

  • Q1. Technical Experience
  • Q2. Challenged faced
  • Q3. Solutions provided
  • Q4. Toughest Project Handled
  • Ans. 

    Developing a real-time monitoring system for a large-scale financial trading platform

    • Implemented complex algorithms to analyze market data in real-time

    • Designed a scalable architecture to handle high volume of transactions

    • Worked closely with traders and analysts to understand requirements and make necessary adjustments

    • Managed a team of developers to ensure project milestones were met

  • Answered by AI
  • Q5. Other Technical languages

Interview Preparation Tips

Interview preparation tips for other job seekers - If they feel you can fill their need then u r in
Interview experience
3
Average
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Coding Test 

Sql, pyhton, azure databricks, azure data factory

Interview experience
5
Excellent
Difficulty level
-
Process Duration
-
Result
-

I applied via Campus Placement

Round 1 - Aptitude Test 

It was communication assesment round.It had good set of questions

Interview experience
5
Excellent
Difficulty level
Easy
Process Duration
Less than 2 weeks
Result
Selected Selected

I applied via Campus Placement and was interviewed in Feb 2023. There were 3 interview rounds.

Round 1 - Resume Shortlist 
Pro Tip by AmbitionBox:
Keep your resume crisp and to the point. A recruiter looks at your resume for an average of 6 seconds, make sure to leave the best impression.
View all tips
Round 2 - Aptitude Test 

Verbal.coding. c programing. Reasoning. Logical questions. Phargraf. Sa

Round 3 - Technical 

(5 Questions)

  • Q1. Oops concept Array Local and global variable Cyber security C structure Operating system
  • Q2. What wifro? Why should hire you? What is your weakness and strength? Wifro founder who? Salary's exception wifro? When campany joined? Percentage 10th and12puc? Backlogs your? Self introduction?
  • Q3. Your insfired person
  • Q4. Self introduction What is wifro?
  • Q5. When joining campany?

Interview Preparation Tips

Interview preparation tips for other job seekers - Apti first clear second tr question then hr.
Your cammuncation is important . and self confindence important
Interview experience
3
Average
Difficulty level
Moderate
Process Duration
2-4 weeks
Result
Selected Selected

I applied via Company Website and was interviewed in Jan 2024. There was 1 interview round.

Round 1 - Coding Test 

To automate REST API calls using ROBOT Framework.

Interview Questionnaire 

1 Question

  • Q1. Questions related to basic testing , project details and roles responsibility
Interview experience
4
Good
Difficulty level
Easy
Process Duration
Less than 2 weeks
Result
Selected Selected

I applied via Approached by Company and was interviewed in Jun 2024. There was 1 interview round.

Round 1 - One-on-one 

(2 Questions)

  • Q1. Java questions, spring boot
  • Q2. Microservices, devops, react js
Interview experience
4
Good
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Selected Selected

I applied via Referral and was interviewed in Aug 2023. There were 2 interview rounds.

Round 1 - Resume Shortlist 
Pro Tip by AmbitionBox:
Don’t add your photo or details such as gender, age, and address in your resume. These details do not add any value.
View all tips
Round 2 - Technical 

(15 Questions)

  • Q1. Introduce your self and Explain Your Project and your Role?
  • Q2. Explain Airflow with its Internal Architecture?
  • Ans. 

    Airflow is a platform to programmatically author, schedule, and monitor workflows.

    • Airflow is written in Python and uses Directed Acyclic Graphs (DAGs) to define workflows.

    • It has a web-based UI for visualization and monitoring of workflows.

    • Airflow consists of a scheduler, a metadata database, a web server, and an executor.

    • Tasks in Airflow are defined as operators, which determine what actually gets executed.

    • Example: A D...

  • Answered by AI
  • Q3. What is RDD in Spark?
  • Ans. 

    RDD stands for Resilient Distributed Dataset in Spark, which is an immutable distributed collection of objects.

    • RDD is the fundamental data structure in Spark, representing a collection of elements that can be operated on in parallel.

    • RDDs are fault-tolerant, meaning they can automatically recover from failures.

    • RDDs support two types of operations: transformations (creating a new RDD from an existing one) and actions (tr

  • Answered by AI
  • Q4. Define RDD Lineage and its Process
  • Ans. 

    RDD Lineage is the record of transformations applied to an RDD and the dependencies between RDDs.

    • RDD Lineage tracks the sequence of transformations applied to an RDD from its source data.

    • It helps in fault tolerance by allowing RDDs to be reconstructed in case of data loss.

    • RDD Lineage is used in Spark to optimize the execution plan by eliminating unnecessary computations.

    • Example: If an RDD is created from a text file an...

  • Answered by AI
  • Q5. What do you mean by broadcast Variables?
  • Ans. 

    Broadcast Variables are read-only shared variables that are cached on each machine in a Spark cluster rather than being sent with tasks.

    • Broadcast Variables are used to efficiently distribute large read-only datasets to all worker nodes in a Spark cluster.

    • They are useful for tasks that require the same data to be shared across multiple stages of a job.

    • Broadcast Variables are created using the broadcast() method in Spark...

  • Answered by AI
  • Q6. What is Broadcasting are you using Broadcasting and what is the limitation of broadcasting?
  • Ans. 

    Broadcasting is a technique used in Apache Spark to optimize data transfer by sending smaller data to all nodes in a cluster.

    • Broadcasting is used to efficiently distribute read-only data to all nodes in a cluster to avoid unnecessary data shuffling.

    • It is commonly used when joining large datasets with smaller lookup tables.

    • Broadcast variables are cached in memory and reused across multiple stages of a Spark job.

    • The limi...

  • Answered by AI
  • Q7. Are you using acumulator and Explain cathelyst optimizer
  • Ans. 

    Accumulators are used for aggregating values across tasks, while Catalyst optimizer is a query optimizer for Apache Spark.

    • Accumulators are variables that are only added to through an associative and commutative operation and can be used to implement counters or sums.

    • Catalyst optimizer is a rule-based query optimizer that leverages advanced programming language features to build an extensible query optimizer.

    • Catalyst op...

  • Answered by AI
  • Q8. Suppose you adding a block and that takes much time you have to debug it how you start the debug ?
  • Ans. 

    To debug a slow block, start by identifying potential bottlenecks, analyzing logs, checking for errors, and profiling the code.

    • Identify potential bottlenecks in the code or system that could be causing the slow performance.

    • Analyze logs and error messages to pinpoint any issues or exceptions that may be occurring.

    • Use profiling tools to analyze the performance of the code and identify areas that need optimization.

    • Check f...

  • Answered by AI
  • Q9. You have to 200 Petabyte of data to load how you will decide the number of executor required ?out of cache you have
  • Ans. 

    The number of executors required to load 200 Petabytes of data depends on the size of each executor and the available cache.

    • Calculate the size of each executor based on available resources and data size

    • Consider the amount of cache available for data processing

    • Determine the optimal number of executors based on the above factors

  • Answered by AI
  • Q10. What is prepartition ?
  • Ans. 

    Prepartition is the process of dividing data into smaller partitions before performing any operations on it.

    • Prepartitioning helps in improving query performance by reducing the amount of data that needs to be processed.

    • It can also help in distributing data evenly across multiple nodes in a distributed system.

    • Examples include partitioning a large dataset based on a specific column like date or region before running anal

  • Answered by AI
  • Q11. Sql Query Table Name Employee column Employee name Salary Department first read this csv file and then write the query in pyspark to find out the name of the employee whose salary is 2nd highest in eac...
  • Q12. Suppose you have string values now you have to find out the frequency of values ? For Example like input ['a' ,'a' ,'a', 'b', 'b', 'c' ] output a,3 b,2 c,1
  • Q13. What is case classes in python ?
  • Ans. 

    Case classes in Python are classes that are used to create immutable objects for pattern matching and data modeling.

    • Case classes are typically used in functional programming to represent data structures.

    • They are immutable, meaning their values cannot be changed once they are created.

    • Case classes automatically define equality, hash code, and toString methods based on the class constructor arguments.

    • They are commonly use...

  • Answered by AI
  • Q14. Suppose there is 100 column in a file i just want to only load 10 column from 100 column how you approach this?
  • Q15. What is lambda Architecture and lambda function?
  • Ans. 

    Lambda Architecture is a data processing architecture designed to handle massive quantities of data by taking advantage of both batch and stream processing methods. Lambda function is a small anonymous function that can take any number of arguments, but can only have one expression.

    • Lambda Architecture combines batch processing and stream processing to handle large amounts of data efficiently.

    • Batch layer stores and proc...

  • Answered by AI

Interview Preparation Tips

Interview preparation tips for other job seekers - Prepare more around Pyspark and SQL

Skills evaluated in this interview

Interview experience
1
Bad
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Resume Shortlist 
Pro Tip by AmbitionBox:
Properly align and format text in your resume. A recruiter will have to spend more time reading poorly aligned text, leading to high chances of rejection.
View all tips
Round 2 - One-on-one 

(4 Questions)

  • Q1. According to smt department
  • Q2. Electronic control unit
  • Q3. Chip mounter, AOI, SPG, SPI, LASER MACHINE
  • Q4. Reflow, conveyor, ok board loder, ng board unloder

Interview Preparation Tips

Interview preparation tips for other job seekers - Very responsibility on working time . In my life part of wipro company. It is very hard working in honesty in my life

Optum Global Solutions Interview FAQs

How many rounds are there in Optum Global Solutions Data Engineering Manager interview?
Optum Global Solutions interview process usually has 2 rounds. The most common rounds in the Optum Global Solutions interview process are Technical and HR.

Tell us how to improve this page.

Optum Global Solutions Data Engineering Manager Interview Process

based on 1 interview

Interview experience

5
  
Excellent
View more

Interview Questions from Similar Companies

TCS Interview Questions
3.7
 • 10.4k Interviews
Accenture Interview Questions
3.9
 • 8.2k Interviews
Infosys Interview Questions
3.6
 • 7.6k Interviews
Wipro Interview Questions
3.7
 • 5.6k Interviews
Cognizant Interview Questions
3.8
 • 5.6k Interviews
Capgemini Interview Questions
3.7
 • 4.8k Interviews
Tech Mahindra Interview Questions
3.5
 • 3.8k Interviews
HCLTech Interview Questions
3.5
 • 3.8k Interviews
Genpact Interview Questions
3.8
 • 3.1k Interviews
LTIMindtree Interview Questions
3.8
 • 3k Interviews
View all
Optum Global Solutions Data Engineering Manager Salary
based on 7 salaries
₹34 L/yr - ₹44 L/yr
8% more than the average Data Engineering Manager Salary in India
View more details

Optum Global Solutions Data Engineering Manager Reviews and Ratings

based on 2 reviews

4.5/5

Rating in categories

3.5

Skill development

4.5

Work-life balance

2.9

Salary

4.5

Job security

4.0

Company culture

2.5

Promotions

4.0

Work satisfaction

Explore 2 Reviews and Ratings
Data Engineering Manager

Gurgaon / Gurugram

6-10 Yrs

Not Disclosed

Explore more jobs
Claims Associate
4.3k salaries
unlock blur

₹1.6 L/yr - ₹5.6 L/yr

Senior Software Engineer
2.8k salaries
unlock blur

₹9.4 L/yr - ₹29.6 L/yr

Software Engineer
2.6k salaries
unlock blur

₹6.2 L/yr - ₹22 L/yr

Senior Claims Associate
1.2k salaries
unlock blur

₹2.1 L/yr - ₹5.8 L/yr

Medical Coder
1.1k salaries
unlock blur

₹1.5 L/yr - ₹8 L/yr

Explore more salaries
Compare Optum Global Solutions with

Cognizant

3.8
Compare

Accenture

3.8
Compare

IBM

4.0
Compare

TCS

3.7
Compare
Did you find this page helpful?
Yes No
write
Share an Interview