Upload Button Icon Add office photos
Premium Employer

i

This company page is being actively managed by Infogain Team. If you also belong to the team, you can get access from here

Infogain Verified Tick

Compare button icon Compare button icon Compare

Filter interviews by

Clear (1)

Infogain Data Engineering Analyst Interview Questions and Answers

Updated 2 Mar 2025

Infogain Data Engineering Analyst Interview Experiences

1 interview found

Interview experience
4
Good
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Assignment 

It's easy to clear and once you clear u will get questioned for interview

Round 2 - One-on-one 

(1 Question)

  • Q1. Basic to High easy and start with High level
Round 3 - HR 

(1 Question)

  • Q1. What is join all use case of join
  • Ans. 

    Join operations in SQL combine data from multiple tables based on related columns, enhancing data analysis capabilities.

    • Inner Join: Returns records with matching values in both tables. Example: Joining 'Employees' and 'Departments' on 'DeptID'.

    • Left Join: Returns all records from the left table and matched records from the right. Example: All 'Employees' with their 'Departments', even if some don't belong to any.

    • Right J...

  • Answered by AI

Interview questions from similar companies

I applied via Walk-in and was interviewed before Feb 2020. There was 1 interview round.

Interview Questionnaire 

1 Question

  • Q1. Interview mainly asked about spark architecture.

Interview Preparation Tips

Interview preparation tips for other job seekers - Prepare well with the basic

Data Scientist Interview Questions & Answers

LTIMindtree user image Abhishek Srivastav

posted on 16 Mar 2015

Interview Questionnaire 

3 Questions

  • Q1. Code For parse Traingle
  • Ans. 

    Code for parsing a triangle

    • Use a loop to iterate through each line of the triangle

    • Split each line into an array of numbers

    • Store the parsed numbers in a 2D array or a list of lists

  • Answered by AI
  • Q2. Asci value along with alphabets(both capital and small)
  • Ans. 

    The ASCII value is a numerical representation of a character. It includes both capital and small alphabets.

    • ASCII values range from 65 to 90 for capital letters A to Z.

    • ASCII values range from 97 to 122 for small letters a to z.

    • For example, the ASCII value of 'A' is 65 and the ASCII value of 'a' is 97.

  • Answered by AI
  • Q3. Would you like to go for Hire studies

Interview Preparation Tips

Round: Test
Experience: First round was through Elitmus.
If you want to be in IT industry must appear it atleast once, for core also u can try it.
It's usually a tough exam but if u are good in maths , apti you will crack it.
Tips: Focus more on Reasoning part. this is the most difficult part.
practise paragraphs reading and solving(Average level)(Infosys level or less)
If you need any kind of help you can contact me via email or can even ring me.
I would recomend everybody to appear this exam with minimum of one month dedicated preparation
Duration: 120 minutes
Total Questions: 60

Round: Coding Round on their own plateform
Experience: It was little difficult to write codes on some other plateform. But time was enough to cope up.
Tips: Try writing as many programs you can write in C, C++ and JAVA not on paper, on compiler . while giving this exam you can select any of these three languages. Based on that your technical Interview will be taken.

Round: Technical Interview
Experience: Its easy one if you have hands on on programing
Tips: Explore and explore .

Round: HR Interview
Experience: Most difficult round for me(I feel myself a little weak in English). But stay calm. And be cheerful.
I still don't know the exact answer of the question but conversation gone for about 20 minutes on this topic.
He din't seem satisfied with me. Btw most of the people says to say no. You can take your call according to the situation.
Tips: Stay calm. Have as much Knowledge about the organisation. Try to make your Intro as much interesting as possible with achivements, hobbies etc. Ya English plays most important role here.

General Tips: Always have faith in yourself. And remember Everything happens for some good reason
Skill Tips: Dont go deep in OS, DBMS but have rough idea about all the topics
Skills: C, C++, DATA STRUCTURE, DBMS, OS
College Name: GANDHI INSTITUTE OF ENGINEERING AND TECHNOLOGY
Motivation: I wanted a job. :)
Funny Moments: A number of stories are there related to this job.
One is I already had an offer so I booked my ticket to home from Bangalore But at very last moment my father told me that you should never miss any chance, go for it. I went and interview date was postponded due to some reasons. I got a mail at 10:30 pm saying I have to attend interview next day morning at 8:30 pm. I ran to get printout of that mail. The venue was 3 hour journey from my place so I din't sleep for the whole night because i knew that if I ll sleep, I would not be able to wake up But I din't studied also because it would have lead to sleep. And Without having sleep and last moment study I made it.

Skills evaluated in this interview

I applied via Campus Placement and was interviewed before Jul 2020. There was 1 interview round.

Interview Questionnaire 

1 Question

  • Q1. Joined as a fresher, basic C program algo

Interview Preparation Tips

Interview preparation tips for other job seekers - Very easy to crack

Interview Questionnaire 

1 Question

  • Q1. What is the architecture of Spark
  • Ans. 

    Spark has a master-slave architecture with a cluster manager and worker nodes.

    • Spark has a driver program that communicates with a cluster manager to allocate resources and schedule tasks.

    • The cluster manager can be standalone, Mesos, or YARN.

    • Worker nodes execute tasks and store data in memory or on disk.

    • Spark can also utilize external data sources like Hadoop Distributed File System (HDFS) or Amazon S3.

    • Spark supports va...

  • Answered by AI

Skills evaluated in this interview

I applied via Recruitment Consultant and was interviewed in Mar 2021. There was 1 interview round.

Interview Questionnaire 

1 Question

  • Q1. Explain about your projects

Interview Preparation Tips

Interview preparation tips for other job seekers - Interviewer was looking for Data science experience in infrastructure that is building a solution for remedy ticket

I applied via Referral and was interviewed in Mar 2021. There were 4 interview rounds.

Interview Questionnaire 

2 Questions

  • Q1. What is data science
  • Ans. 

    Data science is the field of extracting insights and knowledge from data using various techniques and tools.

    • Data science involves collecting, cleaning, and analyzing data to extract insights.

    • It uses various techniques such as machine learning, statistical modeling, and data visualization.

    • Data science is used in various fields such as finance, healthcare, and marketing.

    • Examples of data science applications include fraud...

  • Answered by AI
  • Q2. What is phyton and R
  • Ans. 

    Python and R are programming languages commonly used in data science and statistical analysis.

    • Python is a general-purpose language with a large community and many libraries for data manipulation and machine learning.

    • R is a language specifically designed for statistical computing and graphics, with a wide range of packages for data analysis and visualization.

    • Both languages are popular choices for data scientists and hav...

  • Answered by AI

Interview Preparation Tips

Interview preparation tips for other job seekers - Provide the tips how to face the interview

Skills evaluated in this interview

Interview Questionnaire 

1 Question

  • Q1. Please tell me something about yourself.What is your experience? What are your goals and ambitions?Why We should hire you? Strengths and weaknesses etc.
Interview experience
1
Bad
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Technical 

(1 Question)

  • Q1. All abt sql and databricks Then some questions from adf
Interview experience
4
Good
Difficulty level
Moderate
Process Duration
2-4 weeks
Result
Not Selected

I applied via Naukri.com and was interviewed in Sep 2024. There was 1 interview round.

Round 1 - Technical 

(14 Questions)

  • Q1. How to create pipeline in adf?
  • Ans. 

    To create a pipeline in ADF, you can use the Azure Data Factory UI or code-based approach.

    • Use Azure Data Factory UI to visually create and manage pipelines

    • Use code-based approach with JSON to define pipelines and activities

    • Add activities such as data movement, data transformation, and data processing to the pipeline

    • Set up triggers and schedules for the pipeline to run automatically

  • Answered by AI
  • Q2. Diffrent types of activities in pipelines
  • Ans. 

    Activities in pipelines include data extraction, transformation, loading, and monitoring.

    • Data extraction: Retrieving data from various sources such as databases, APIs, and files.

    • Data transformation: Cleaning, filtering, and structuring data for analysis.

    • Data loading: Loading processed data into a data warehouse or database.

    • Monitoring: Tracking the performance and health of the pipeline to ensure data quality and reliab

  • Answered by AI
  • Q3. What is use of getmetadata
  • Ans. 

    getmetadata is used to retrieve metadata information about a dataset or data source.

    • getmetadata can provide information about the structure, format, and properties of the data.

    • It can be used to understand the data schema, column names, data types, and any constraints or relationships.

    • This information is helpful for data engineers to properly process, transform, and analyze the data.

    • For example, getmetadata can be used ...

  • Answered by AI
  • Q4. Diffrent types of triggers
  • Ans. 

    Triggers in databases are special stored procedures that are automatically executed when certain events occur.

    • Types of triggers include: DML triggers (for INSERT, UPDATE, DELETE operations), DDL triggers (for CREATE, ALTER, DROP operations), and logon triggers.

    • Triggers can be classified as row-level triggers (executed once for each row affected by the triggering event) or statement-level triggers (executed once for eac...

  • Answered by AI
  • Q5. Diffrence between normal cluster and job cluster in databricks
  • Ans. 

    Normal cluster is used for interactive workloads while job cluster is used for batch processing in Databricks.

    • Normal cluster is used for ad-hoc queries and exploratory data analysis.

    • Job cluster is used for running scheduled jobs and batch processing tasks.

    • Normal cluster is terminated after a period of inactivity, while job cluster is terminated after the job completes.

    • Normal cluster is more cost-effective for short-liv...

  • Answered by AI
  • Q6. What is slowly changing dimensions
  • Ans. 

    Slowly changing dimensions refer to data warehouse dimensions that change slowly over time.

    • SCDs are used to track historical changes in data over time.

    • There are three types of SCDs - Type 1, Type 2, and Type 3.

    • Type 1 SCDs overwrite old data with new data, Type 2 creates new records for changes, and Type 3 maintains both old and new data in separate columns.

    • Example: A customer's address changing would be a Type 2 SCD.

    • Ex...

  • Answered by AI
  • Q7. Incremental load
  • Q8. With use in python
  • Ans. 

    Use Python's 'with' statement to ensure proper resource management and exception handling.

    • Use 'with' statement to automatically close files after use

    • Helps in managing resources like database connections

    • Ensures proper cleanup even in case of exceptions

  • Answered by AI
  • Q9. List vs tuple in python
  • Ans. 

    List is mutable, tuple is immutable in Python.

    • List can be modified after creation, tuple cannot be modified.

    • List uses square brackets [], tuple uses parentheses ().

    • Lists are used for collections of items that may need to be changed, tuples are used for fixed collections of items.

    • Example: list_example = [1, 2, 3], tuple_example = (4, 5, 6)

  • Answered by AI
  • Q10. Datalake 1 vs datalake2
  • Ans. 

    Datalake 1 and Datalake 2 are both storage systems for big data, but they may differ in terms of architecture, scalability, and use cases.

    • Datalake 1 may use a Hadoop-based architecture while Datalake 2 may use a cloud-based architecture like AWS S3 or Azure Data Lake Storage.

    • Datalake 1 may be more suitable for on-premise data storage and processing, while Datalake 2 may offer better scalability and flexibility for clou...

  • Answered by AI
  • Q11. How to read a file in databricks
  • Ans. 

    To read a file in Databricks, you can use the Databricks File System (DBFS) or Spark APIs.

    • Use dbutils.fs.ls('dbfs:/path/to/file') to list files in DBFS

    • Use spark.read.format('csv').load('dbfs:/path/to/file') to read a CSV file

    • Use spark.read.format('parquet').load('dbfs:/path/to/file') to read a Parquet file

  • Answered by AI
  • Q12. Star vs snowflake schema
  • Ans. 

    Star schema is denormalized with one central fact table surrounded by dimension tables, while snowflake schema is normalized with multiple related dimension tables.

    • Star schema is easier to understand and query due to denormalization.

    • Snowflake schema saves storage space by normalizing data.

    • Star schema is better for data warehousing and OLAP applications.

    • Snowflake schema is better for OLTP systems with complex relationsh

  • Answered by AI
  • Q13. Repartition vs coalesece
  • Ans. 

    repartition increases partitions while coalesce decreases partitions in Spark

    • repartition shuffles data and can be used for increasing partitions for parallelism

    • coalesce reduces partitions without shuffling data, useful for reducing overhead

    • repartition is more expensive than coalesce as it involves data movement

    • example: df.repartition(10) vs df.coalesce(5)

  • Answered by AI
  • Q14. Parquet file uses
  • Ans. 

    Parquet file format is a columnar storage format used for efficient data storage and processing.

    • Parquet files store data in a columnar format, which allows for efficient querying and processing of specific columns without reading the entire file.

    • It supports complex nested data structures like arrays and maps.

    • Parquet files are highly compressed, reducing storage space and improving query performance.

    • It is commonly used ...

  • Answered by AI

Skills evaluated in this interview

Contribute & help others!
anonymous
You can choose to be anonymous

Infogain Interview FAQs

How many rounds are there in Infogain Data Engineering Analyst interview?
Infogain interview process usually has 3 rounds. The most common rounds in the Infogain interview process are Assignment, One-on-one Round and HR.
What are the top questions asked in Infogain Data Engineering Analyst interview?

Some of the top questions asked at the Infogain Data Engineering Analyst interview -

  1. What is join all use case of j...read more
  2. Basic to High easy and start with High le...read more

Recently Viewed

INTERVIEWS

Infogain

No Interviews

INTERVIEWS

CommScope

No Interviews

INTERVIEWS

Infogain

No Interviews

INTERVIEWS

KEC International

No Interviews

SALARIES

Infogain

SALARIES

Infogain

SALARIES

Infogain

INTERVIEWS

Diversitech General Engineering

No Interviews

INTERVIEWS

Infosys

No Interviews

SALARIES

Vindhya Telelinks

Tell us how to improve this page.

Infogain Data Engineering Analyst Interview Process

based on 1 interview

Interview experience

4
  
Good
View more

Interview Questions from Similar Companies

TCS Interview Questions
3.7
 • 10.4k Interviews
Infosys Interview Questions
3.6
 • 7.5k Interviews
Wipro Interview Questions
3.7
 • 5.6k Interviews
Tech Mahindra Interview Questions
3.5
 • 3.8k Interviews
HCLTech Interview Questions
3.5
 • 3.8k Interviews
LTIMindtree Interview Questions
3.8
 • 2.9k Interviews
Mphasis Interview Questions
3.4
 • 790 Interviews
KPIT Technologies Interview Questions
3.4
 • 280 Interviews
CitiusTech Interview Questions
3.4
 • 269 Interviews
View all
Software Engineer
1.3k salaries
unlock blur

₹0 L/yr - ₹0 L/yr

Development Consultant
811 salaries
unlock blur

₹0 L/yr - ₹0 L/yr

Technical Lead
613 salaries
unlock blur

₹0 L/yr - ₹0 L/yr

Consultant
458 salaries
unlock blur

₹0 L/yr - ₹0 L/yr

Senior Consultant
373 salaries
unlock blur

₹0 L/yr - ₹0 L/yr

Explore more salaries
Compare Infogain with

TCS

3.7
Compare

Wipro

3.7
Compare

HCLTech

3.5
Compare

Tech Mahindra

3.5
Compare
Did you find this page helpful?
Yes No
write
Share an Interview
Rate your experience using AmbitionBox
Terrible
Terrible
Poor
Poor
Average
Average
Good
Good
Excellent
Excellent