Upload Button Icon Add office photos
Engaged Employer

i

This company page is being actively managed by Mindgraph Team. If you also belong to the team, you can get access from here

Mindgraph Verified Tick

Compare button icon Compare button icon Compare
3.4

based on 49 Reviews

Filter interviews by

Mindgraph Data Engineer Interview Questions and Answers

Updated 8 Oct 2024

Mindgraph Data Engineer Interview Experiences

1 interview found

Interview experience
4
Good
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Selected Selected

I applied via campus placement at Gayatri Vidya Parishad College of Engineering, Visakhapatnam and was interviewed before Oct 2023. There were 3 interview rounds.

Round 1 - Aptitude Test 

Aptitude, reasoning & coding

Round 2 - Case Study 

Groupwise we have to complete task for data engineering, and need to explain what we have done.

Round 3 - One-on-one 

(4 Questions)

  • Q1. Introduce yourself
  • Ans. 

    I am a data engineer with a strong background in programming and data analysis.

    • Experienced in programming languages such as Python, SQL, and Java

    • Skilled in data manipulation, ETL processes, and data modeling

    • Familiar with big data technologies like Hadoop, Spark, and Kafka

  • Answered by AI
  • Q2. In-depth questions about my final year project
  • Q3. Given a condition asked to write a python code for that on board
  • Q4. Some basic data science questions like k-means, Naive bayesian etc

Interview Preparation Tips

Interview preparation tips for other job seekers - more focus on python and sql when applying for data Engineer or Analyst roles

Interview questions from similar companies

Interview experience
3
Average
Difficulty level
-
Process Duration
-
Result
No response
Round 1 - Technical 

(4 Questions)

  • Q1. What is the architecture of Apache Spark?
  • Ans. 

    Apache Spark architecture includes a cluster manager, worker nodes, and driver program.

    • Apache Spark architecture consists of a cluster manager, which allocates resources and schedules tasks.

    • Worker nodes execute tasks and store data in memory or disk.

    • Driver program coordinates tasks and communicates with the cluster manager.

    • Spark applications run as independent sets of processes on a cluster, coordinated by the SparkCon...

  • Answered by AI
  • Q2. What is the difference between the reduceBy and groupBy transformations in Apache Spark?
  • Ans. 

    reduceBy is used to aggregate data based on key, while groupBy is used to group data based on key.

    • reduceBy is a transformation that combines the values of each key using an associative function and a neutral 'zero value'.

    • groupBy is a transformation that groups the data based on a key and returns a grouped data set.

    • reduceBy is more efficient for aggregating data as it reduces the data before shuffling, while groupBy shu...

  • Answered by AI
  • Q3. What is the difference between RDD (Resilient Distributed Datasets) and DataFrame in Apache Spark?
  • Ans. 

    RDD is a low-level abstraction representing a distributed collection of objects, while DataFrame is a higher-level abstraction representing a distributed collection of data organized into named columns.

    • RDD is more suitable for unstructured data and low-level transformations, while DataFrame is more suitable for structured data and high-level abstractions.

    • DataFrames provide optimizations like query optimization and code...

  • Answered by AI
  • Q4. What are the different modes of execution in Apache Spark?
  • Ans. 

    The different modes of execution in Apache Spark include local mode, standalone mode, YARN mode, and Mesos mode.

    • Local mode: Spark runs on a single machine with one executor.

    • Standalone mode: Spark runs on a cluster managed by a standalone cluster manager.

    • YARN mode: Spark runs on a Hadoop cluster using YARN as the resource manager.

    • Mesos mode: Spark runs on a Mesos cluster with Mesos as the resource manager.

  • Answered by AI
Interview experience
5
Excellent
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Not Selected

I applied via Naukri.com and was interviewed in Oct 2024. There were 2 interview rounds.

Round 1 - One-on-one 

(2 Questions)

  • Q1. Azure Scenario based questions
  • Q2. Pyspark Coding based questions
Round 2 - One-on-one 

(2 Questions)

  • Q1. ADF, Databricks related question
  • Q2. Spark Performance problem and scenarios
Interview experience
2
Poor
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
No response

I applied via Campus Placement and was interviewed in Oct 2024. There was 1 interview round.

Round 1 - Technical 

(2 Questions)

  • Q1. Write code in regular expression to remove the special characters
  • Ans. 

    Use regular expression to remove special characters from a string

    • Use the regex pattern [^a-zA-Z0-9\s] to match any character that is not a letter, digit, or whitespace

    • Use the replace() function in your programming language to replace the matched special characters with an empty string

    • Example: input string 'Hello! How are you?' will become 'Hello How are you' after removing special characters

  • Answered by AI
  • Q2. Questions on resume
Interview experience
4
Good
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Technical 

(2 Questions)

  • Q1. What is Databricks, Internals of it, Optimization technique
  • Ans. 

    Databricks is a unified analytics platform that provides a collaborative environment for data scientists, engineers, and analysts.

    • Databricks is built on top of Apache Spark, providing a unified platform for data engineering, data science, and business analytics.

    • Internals of Databricks include a cluster manager, job scheduler, and workspace for collaboration.

    • Optimization techniques in Databricks include query optimizati...

  • Answered by AI
  • Q2. SQL questions of joins,group by
Round 2 - Technical 

(2 Questions)

  • Q1. Scenario based azure data factory questions
  • Q2. Project structres, pyspark dataframes related

Interview Preparation Tips

Interview preparation tips for other job seekers - Juat practice fundamentals

Skills evaluated in this interview

Interview experience
3
Average
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Not Selected

I applied via Company Website and was interviewed in Jul 2024. There was 1 interview round.

Round 1 - Technical 

(2 Questions)

  • Q1. Python Lambda Function
  • Q2. What are pods in Kubernetes
  • Ans. 

    Pods are the smallest deployable units in Kubernetes, consisting of one or more containers.

    • Pods are used to run and manage containers in Kubernetes

    • Each pod has its own unique IP address within the Kubernetes cluster

    • Pods can contain multiple containers that share resources and are scheduled together

    • Pods are ephemeral and can be easily created, destroyed, or replicated

    • Pods can be managed and scaled using Kubernetes contr

  • Answered by AI

Skills evaluated in this interview

Interview experience
2
Poor
Difficulty level
-
Process Duration
Less than 2 weeks
Result
Not Selected

I applied via Approached by Company and was interviewed in Apr 2024. There was 1 interview round.

Round 1 - One-on-one 

(3 Questions)

  • Q1. What Volume of data have you handled in your POCs ?
  • Ans. 

    I have handled terabytes of data in my POCs, including data from various sources and formats.

    • Handled terabytes of data in POCs

    • Worked with data from various sources and formats

    • Used tools like Hadoop, Spark, and SQL for data processing

  • Answered by AI
  • Q2. When will you decide to use repartition and coalesce?
  • Ans. 

    Repartition is used for increasing partitions for parallelism, while coalesce is used for decreasing partitions to reduce shuffling.

    • Repartition is used when there is a need for more partitions to increase parallelism.

    • Coalesce is used when there are too many partitions and need to reduce them to avoid shuffling.

    • Example: Repartition can be used before a join operation to evenly distribute data across partitions for bette...

  • Answered by AI
  • Q3. How will you design/configure a cluster if you have given 10 petabytes of data.
  • Ans. 

    Designing/configuring a cluster for 10 petabytes of data involves considerations for storage capacity, processing power, network bandwidth, and fault tolerance.

    • Consider using a distributed file system like HDFS or object storage like Amazon S3 to store and manage the large volume of data.

    • Implement a scalable processing framework like Apache Spark or Hadoop to efficiently process and analyze the data in parallel.

    • Utilize...

  • Answered by AI

Interview Preparation Tips

Interview preparation tips for other job seekers - Sometimes you can easily detect that they want to rush your interview by asking irrelevant questions even though you have not mentioned such expertise in your resume. Also they have only shortlisted the resume just to reject. This breaks the confidence of the person who is looking for the job and honestly conveying the level of expertise he/she has.

Skills evaluated in this interview

Interview experience
5
Excellent
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Selected Selected

I applied via Job Portal and was interviewed in Aug 2024. There were 3 interview rounds.

Round 1 - Aptitude Test 

Its mandatory test even for experience people

Round 2 - Technical 

(1 Question)

  • Q1. Related to technology
Round 3 - HR 

(1 Question)

  • Q1. Very good discussion towards work culture, salary and all
Interview experience
4
Good
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Technical 

(2 Questions)

  • Q1. Data Structures
  • Q2. Python Lambda Function
Round 2 - HR 

(2 Questions)

  • Q1. What are your Strengths
  • Ans. 

    My strengths include strong analytical skills, attention to detail, and problem-solving abilities.

    • Strong analytical skills - able to analyze complex data sets and derive meaningful insights

    • Attention to detail - meticulous in ensuring data accuracy and quality

    • Problem-solving abilities - adept at identifying and resolving data-related issues

    • Experience with data manipulation tools like SQL, Python, and Spark

  • Answered by AI
  • Q2. Why you are looking for job change
  • Ans. 

    Seeking new challenges and growth opportunities in a different environment.

    • Looking for new challenges to enhance my skills and knowledge

    • Seeking growth opportunities that align with my career goals

    • Interested in exploring different technologies and industries

    • Want to work in a more collaborative team environment

    • Seeking better work-life balance or location proximity

  • Answered by AI
Interview experience
4
Good
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Technical 

(2 Questions)

  • Q1. Sql questions were
  • Q2. Spark related aswell

Mindgraph Interview FAQs

How many rounds are there in Mindgraph Data Engineer interview?
Mindgraph interview process usually has 3 rounds. The most common rounds in the Mindgraph interview process are Aptitude Test, Case Study and One-on-one Round.
What are the top questions asked in Mindgraph Data Engineer interview?

Some of the top questions asked at the Mindgraph Data Engineer interview -

  1. Given a condition asked to write a python code for that on bo...read more
  2. Some basic data science questions like k-means, Naive bayesian ...read more

Tell us how to improve this page.

Mindgraph Data Engineer Interview Process

based on 1 interview

Interview experience

4
  
Good
View more
Mindgraph Data Engineer Salary
based on 7 salaries
₹3.6 L/yr - ₹5 L/yr
57% less than the average Data Engineer Salary in India
View more details

Mindgraph Data Engineer Reviews and Ratings

based on 4 reviews

4.7/5

Rating in categories

5.0

Skill development

4.2

Work-life balance

4.5

Salary

4.7

Job security

4.5

Company culture

4.5

Promotions

4.7

Work satisfaction

Explore 4 Reviews and Ratings
Software Engineer
32 salaries
unlock blur

₹2.5 L/yr - ₹11 L/yr

Software Developer
11 salaries
unlock blur

₹3 L/yr - ₹12 L/yr

Azure DevOps Engineer
8 salaries
unlock blur

₹5 L/yr - ₹6.1 L/yr

Data Engineer
7 salaries
unlock blur

₹3.6 L/yr - ₹5 L/yr

Java Developer
7 salaries
unlock blur

₹2.6 L/yr - ₹8.5 L/yr

Explore more salaries
Compare Mindgraph with

TCS

3.7
Compare

Infosys

3.6
Compare

Wipro

3.7
Compare

HCLTech

3.5
Compare
Did you find this page helpful?
Yes No
write
Share an Interview