Upload Button Icon Add office photos
Engaged Employer

i

This company page is being actively managed by Infovision Team. If you also belong to the team, you can get access from here

Infovision Verified Tick

Compare button icon Compare button icon Compare
3.4

based on 270 Reviews

Filter interviews by

Infovision Data Engineer Interview Questions, Process, and Tips

Updated 21 Feb 2024

Infovision Data Engineer Interview Experiences

1 interview found

Data Engineer Interview Questions & Answers

user image Anonymous

posted on 21 Feb 2024

Interview experience
3
Average
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Not Selected

I was interviewed in Jan 2024.

Round 1 - Technical 

(15 Questions)

  • Q1. What are window functions in SQL
  • Ans. 

    Window functions in SQL are used to perform calculations across a set of table rows related to the current row.

    • Window functions are used to calculate values based on a set of rows related to the current row.

    • They allow for ranking, aggregation, and other calculations within a specific window of rows.

    • Common window functions include ROW_NUMBER(), RANK(), DENSE_RANK(), and SUM() OVER().

  • Answered by AI
  • Q2. ETL - How to do full load in SSIS, mention the steps
  • Ans. 

    To perform a full load in SSIS, you can use the Data Flow Task with a source and destination component.

    • Create a Data Flow Task in the Control Flow tab of the SSIS package.

    • Add a source component to extract data from the source system.

    • Add a destination component to load data into the destination system.

    • Map the columns from the source to the destination.

    • Run the package to execute the full load.

  • Answered by AI
  • Q3. ETL- how to do the incremental load in ADF and in SSIS
  • Ans. 

    Incremental load in ADF and SSIS involves identifying new or updated data and loading only those changes.

    • In ADF, use watermark columns to track the last loaded value and filter data based on this value

    • In SSIS, use CDC (Change Data Capture) components or custom scripts to identify new or updated data

    • Both ADF and SSIS support incremental loading by comparing source and target data to determine changes

  • Answered by AI
  • Q4. Linked Service Vs Dataset
  • Ans. 

    Linked Service connects to external data sources, while Dataset represents the data within the data store.

    • Linked Service is used to connect to external data sources like databases, APIs, and file systems.

    • Dataset represents the data within the data store and can be used for data processing and analysis.

    • Linked Service defines the connection information and credentials needed to access external data sources.

    • Dataset define...

  • Answered by AI
  • Q5. What is IR - integration Runtime? what are the types of IR
  • Ans. 

    Integration Runtime (IR) is a compute infrastructure that provides data integration capabilities across different network environments.

    • IR is used in Azure Data Factory to provide data integration capabilities

    • There are three types of IR: Azure, Self-hosted, and Azure-SSIS

    • Azure IR is fully managed by Microsoft and is used for data movement in the cloud

    • Self-hosted IR allows data movement between on-premises and cloud data...

  • Answered by AI
  • Q6. Scenario: In ADF, if we have files in FTP folder, how to get it to ADLS where the file size > 10mb.
  • Q7. How to copy data without using multiple activities. Dynamically using loops/ parameterization.
  • Ans. 

    Use a single activity with dynamic parameterization and loops to copy data.

    • Use a loop to iterate through the data source and destination locations.

    • Parameterize the source and destination locations to dynamically copy data.

    • Utilize a scripting language like Python or PowerShell to implement the logic.

    • Example: Use a Python script with a loop to copy files from one folder to another.

    • Example: Use PowerShell script with dyna...

  • Answered by AI
  • Q8. Datawarehouse - What is a Fact and Dimention table
  • Q9. What is snowflake?
  • Ans. 

    Snowflake is a cloud-based data warehousing platform that allows for easy and scalable data storage and analysis.

    • Snowflake is a fully managed service that works on a pay-as-you-go model.

    • It separates storage and compute resources, allowing for better scalability and cost-effectiveness.

    • Snowflake supports SQL queries and has built-in support for semi-structured data like JSON and XML.

    • It provides features like automatic sc...

  • Answered by AI
  • Q10. Agile - Scrum team size?
  • Q11. SQL - (Merge) Insert / update data based on the incoming data.
  • Q12. SQL - Delete Vs Truncate ?
  • Ans. 

    Delete removes rows one by one and can be rolled back, while Truncate removes all rows at once and cannot be rolled back.

    • Delete is a DML command, while Truncate is a DDL command.

    • Delete can be rolled back using a transaction, while Truncate cannot be rolled back.

    • Delete fires triggers on each row deletion, while Truncate does not fire triggers.

    • Delete is slower as it removes rows one by one, while Truncate is faster as it...

  • Answered by AI
  • Q13. How to insert non-duplicate data into target table. how many ways we can do.
  • Ans. 

    To insert non-duplicate data into a target table, you can use methods like using a unique constraint, using a merge statement, or using a temporary table.

    • Use a unique constraint on the target table to prevent duplicate entries.

    • Use a merge statement to insert data into the target table only if it does not already exist.

    • Use a temporary table to store the new data, then insert only the non-duplicate records into the targe

  • Answered by AI
  • Q14. SQL - how do you identify the long running queries.
  • Ans. 

    Identifying long running queries in SQL

    • Monitor query execution times using tools like SQL Server Profiler or Performance Monitor

    • Check system views like sys.dm_exec_requests or sys.dm_exec_query_stats for query durations

    • Use query hints like OPTION (RECOMPILE) to force recompilation of queries for better performance

  • Answered by AI
  • Q15. SQL - online test given with 3 queries. One is related to Joins , second one is related to Agg functions, and 3rd is related to LAG function.

Skills evaluated in this interview

Interview questions from similar companies

Interview experience
5
Excellent
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Selected Selected

I was interviewed in Oct 2024.

Round 1 - Technical 

(2 Questions)

  • Q1. SQL problem window function
  • Q2. SQL code like join and scenerio
Round 2 - Technical 

(2 Questions)

  • Q1. Design round for adf pipeline
  • Ans. 

    Designing an ADF pipeline for data processing

    • Identify data sources and destinations

    • Define data transformations and processing steps

    • Consider scheduling and monitoring requirements

    • Utilize ADF activities like Copy Data, Data Flow, and Databricks

    • Implement error handling and logging mechanisms

  • Answered by AI
  • Q2. Azure synapses and adf adb
Round 3 - HR 

(2 Questions)

  • Q1. Expected ctc and current ctc negotiations
  • Ans. 

    Discussing expected and current salary for negotiation purposes.

    • Be honest about your current salary and provide a realistic expectation for your desired salary.

    • Highlight your skills and experience that justify your desired salary.

    • Be open to negotiation and willing to discuss other benefits besides salary.

    • Research industry standards and salary ranges for similar positions to support your negotiation.

    • Focus on the value y...

  • Answered by AI
  • Q2. Relocation and remote work until ofc open for Pune location

Interview Preparation Tips

Interview preparation tips for other job seekers - Be prepared for sql and problem solving

Skills evaluated in this interview

Interview experience
5
Excellent
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Not Selected

I applied via Naukri.com and was interviewed in Oct 2024. There were 2 interview rounds.

Round 1 - One-on-one 

(2 Questions)

  • Q1. Azure Scenario based questions
  • Q2. Pyspark Coding based questions
Round 2 - One-on-one 

(2 Questions)

  • Q1. ADF, Databricks related question
  • Q2. Spark Performance problem and scenarios
Interview experience
5
Excellent
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Selected Selected

I applied via Recruitment Consulltant and was interviewed in Nov 2024. There were 2 interview rounds.

Round 1 - Technical 

(2 Questions)

  • Q1. Question based on big data
  • Q2. Cloud based messaging system
Round 2 - Technical 

(1 Question)

  • Q1. Cloud based operations
Interview experience
5
Excellent
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Technical 

(2 Questions)

  • Q1. SCD questions. Iceberg questions
  • Q2. Basic python programing, pyspark architechture.
Interview experience
4
Good
Difficulty level
-
Process Duration
-
Result
-
Round 1 - One-on-one 

(1 Question)

  • Q1. Spark basic question , hive related questions.

Interview Preparation Tips

Interview preparation tips for other job seekers - Good question asked, It covers sql , spark and python.
Interview experience
5
Excellent
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Aptitude Test 

Mostly verbal and maths questions.

Round 2 - Technical 

(2 Questions)

  • Q1. What is DDL, DML?
  • Ans. 

    DDL stands for Data Definition Language and is used to define the structure of database objects. DML stands for Data Manipulation Language and is used to manipulate data within the database.

    • DDL is used to create, modify, and delete database objects such as tables, indexes, and views

    • DML is used to insert, update, delete, and retrieve data from the database

    • Examples of DDL statements include CREATE TABLE, ALTER TABLE, DRO...

  • Answered by AI
  • Q2. SQL queries
Round 3 - HR 

(2 Questions)

  • Q1. Strength and weakness
  • Q2. Why TTN
Interview experience
5
Excellent
Difficulty level
Moderate
Process Duration
-
Result
Selected Selected
Round 1 - Coding Test 

1- sql adn 10 spark questions.

Round 2 - Technical 

(2 Questions)

  • Q1. Python programs - 4
  • Q2. Sql questions - 4
Round 3 - Technical 

(1 Question)

  • Q1. Project related and work experience questions
Round 4 - HR 

(1 Question)

  • Q1. Managerial question
Interview experience
4
Good
Difficulty level
-
Process Duration
-
Result
-

I applied via Campus Placement

Round 1 - Aptitude Test 

Aptitude test had few quant and verbal questions then SQL MCQs and 3 Coding question

Round 2 - Technical 

(2 Questions)

  • Q1. Questions about guesstimates and puzzles
  • Q2. Which coding language is better and why?
  • Ans. 

    There is no one 'better' coding language, as it depends on the specific use case and requirements.

    • The best coding language depends on the project requirements, team expertise, and ecosystem support.

    • For data engineering, languages like Python, Scala, and SQL are commonly used for their data processing capabilities.

    • Python is popular for its simplicity and extensive libraries like Pandas and NumPy, while Scala is known fo...

  • Answered by AI
Interview experience
3
Average
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Not Selected

I applied via Company Website and was interviewed in Jul 2024. There was 1 interview round.

Round 1 - Technical 

(2 Questions)

  • Q1. Python Lambda Function
  • Q2. What are pods in Kubernetes
  • Ans. 

    Pods are the smallest deployable units in Kubernetes, consisting of one or more containers.

    • Pods are used to run and manage containers in Kubernetes

    • Each pod has its own unique IP address within the Kubernetes cluster

    • Pods can contain multiple containers that share resources and are scheduled together

    • Pods are ephemeral and can be easily created, destroyed, or replicated

    • Pods can be managed and scaled using Kubernetes contr

  • Answered by AI

Skills evaluated in this interview

Infovision Interview FAQs

How many rounds are there in Infovision Data Engineer interview?
Infovision interview process usually has 1 rounds. The most common rounds in the Infovision interview process are Technical.
How to prepare for Infovision Data Engineer interview?
Go through your CV in detail and study all the technologies mentioned in your CV. Prepare at least two technologies or languages in depth if you are appearing for a technical interview at Infovision. The most common topics and skills that interviewers at Infovision expect are Airflow, Azure DevOps, Big Data, Data Bricks and Data Engineering.
What are the top questions asked in Infovision Data Engineer interview?

Some of the top questions asked at the Infovision Data Engineer interview -

  1. What is IR - integration Runtime? what are the types of...read more
  2. ETL- how to do the incremental load in ADF and in S...read more
  3. how to insert non-duplicate data into target table. how many ways we can ...read more

Tell us how to improve this page.

Infovision Data Engineer Interview Process

based on 1 interview

Interview experience

3
  
Average
View more
Infovision Data Engineer Salary
based on 6 salaries
₹4.6 L/yr - ₹17.8 L/yr
At par with the average Data Engineer Salary in India
View more details
Softwaretest Engineer
309 salaries
unlock blur

₹2.5 L/yr - ₹6.2 L/yr

Senior Software Engineer
291 salaries
unlock blur

₹6 L/yr - ₹21.7 L/yr

Software Engineer
245 salaries
unlock blur

₹2.8 L/yr - ₹10 L/yr

Technical Lead
176 salaries
unlock blur

₹10.3 L/yr - ₹35 L/yr

Software Developer
143 salaries
unlock blur

₹3 L/yr - ₹10.2 L/yr

Explore more salaries
Compare Infovision with

TCS

3.7
Compare

Wipro

3.7
Compare

HCLTech

3.5
Compare

Tech Mahindra

3.5
Compare
Did you find this page helpful?
Yes No
write
Share an Interview