Upload Button Icon Add office photos

Filter interviews by

Infoqort Technologies Senior Data Engineer Interview Questions and Answers

Updated 17 Oct 2023

Infoqort Technologies Senior Data Engineer Interview Experiences

1 interview found

Interview experience
1
Bad
Difficulty level
-
Process Duration
-
Result
-
Round 1 - One-on-one 

(2 Questions)

  • Q1. Nothing njdfhjjbhjhgfji
  • Q2. Nothing hhhhhhhjjjjjj
Round 2 - One-on-one 

(2 Questions)

  • Q1. Fghjnnkkknvccccggbhhh
  • Q2. Gggghuujjkkjjhhhhhh

Interview Preparation Tips

Interview preparation tips for other job seekers - Don't go for this company.worst company.i have ever worked with.everyone from owner to HR debosree.all are liar.you will not get salary.CEO Sanjeev sadhwani, project manager head harpreet singh and HR Debosree all are liar.if you want to play with your life then only choose this company.

Interview questions from similar companies

Interview experience
5
Excellent
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
-

I applied via Recruitment Consulltant

Round 1 - Technical 

(5 Questions)

  • Q1. Explain ETL pipeline ecosystem in Azure Databricks?
  • Q2. Star vs Snowflake schema, when to use?
  • Q3. Find Salary higher than Average department salary
  • Q4. Implementation of SCD2 table
  • Q5. How incremental loading is done
Interview experience
4
Good
Difficulty level
Easy
Process Duration
Less than 2 weeks
Result
No response

I applied via Naukri.com and was interviewed in Nov 2024. There was 1 interview round.

Round 1 - Technical 

(6 Questions)

  • Q1. Can you introduce yourself and describe your current project experience?
  • Ans. 

    I am a Senior Data Engineer with experience in building scalable data pipelines and optimizing data processing workflows.

    • Experience in designing and implementing ETL processes using tools like Apache Spark and Airflow

    • Proficient in working with large datasets and optimizing query performance

    • Strong background in data modeling and database design

    • Worked on projects involving real-time data processing and streaming analytic

  • Answered by AI
  • Q2. Decorators in python
  • Ans. 

    Decorators in Python are functions that modify the behavior of other functions or methods.

    • Decorators are defined using the @decorator_name syntax before a function definition.

    • They can be used to add functionality to existing functions without modifying their code.

    • Decorators can be used for logging, timing, authentication, and more.

    • Example: @staticmethod decorator in Python is used to define a static method in a class.

  • Answered by AI
  • Q3. What is the SQL query to group by employee ID in order to combine the first name and last name with a space?
  • Ans. 

    SQL query to group by employee ID and combine first name and last name with a space

    • Use the GROUP BY clause to group by employee ID

    • Use the CONCAT function to combine first name and last name with a space

    • Select employee ID, CONCAT(first_name, ' ', last_name) AS full_name

  • Answered by AI
  • Q4. What are constructors in Python?
  • Ans. 

    Constructors in Python are special methods used for initializing objects. They are called automatically when a new instance of a class is created.

    • Constructors are defined using the __init__() method in a class.

    • They are used to initialize instance variables of a class.

    • Example: class Person: def __init__(self, name, age): self.name = name self.age = age person1 = Person('Alice', 30)

  • Answered by AI
  • Q5. Indexing in sql
  • Ans. 

    Indexing in SQL is a technique used to improve the performance of queries by creating a data structure that allows for faster retrieval of data.

    • Indexes are created on columns in a database table to speed up the retrieval of rows that match a certain condition in a WHERE clause.

    • Indexes can be created using CREATE INDEX statement in SQL.

    • Types of indexes include clustered indexes, non-clustered indexes, unique indexes, an...

  • Answered by AI
  • Q6. Why spark works well with parquet files?
  • Ans. 

    Spark works well with Parquet files due to its columnar storage format, efficient compression, and ability to push down filters.

    • Parquet files are columnar storage format, which aligns well with Spark's processing model of working on columns rather than rows.

    • Parquet files support efficient compression, reducing storage space and improving read performance in Spark.

    • Spark can push down filters to Parquet files, allowing f...

  • Answered by AI

Skills evaluated in this interview

Interview experience
5
Excellent
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Selected Selected

I was interviewed in Aug 2024.

Round 1 - Coding Test 

Python and sql tasks

Round 2 - Technical 

(2 Questions)

  • Q1. Project related questions
  • Q2. Coding questions on pyspark windows
Round 3 - One-on-one 

(1 Question)

  • Q1. Managerial discussions, mostly in and around the previous projects
Interview experience
4
Good
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
No response

I was interviewed in Sep 2024.

Round 1 - Technical 

(3 Questions)

  • Q1. Coding in pyspark
  • Ans. 

    Pyspark is a Python library for big data processing using Spark framework.

    • Pyspark is used for processing large datasets in parallel.

    • It provides APIs for data manipulation, querying, and analysis.

    • Example: Using pyspark to read a CSV file and perform data transformations.

  • Answered by AI
  • Q2. Databricks optimisation technique
  • Ans. 

    Databricks optimisation techniques improve performance and efficiency of data processing on the Databricks platform.

    • Use cluster sizing and autoscaling to optimize resource allocation based on workload

    • Leverage Databricks Delta for optimized data storage and processing

    • Utilize caching and persisting data to reduce computation time

    • Optimize queries by using appropriate indexing and partitioning strategies

  • Answered by AI
  • Q3. Aqe details in databricks
  • Ans. 

    Databricks is a unified data analytics platform that provides a collaborative environment for data engineers.

    • Databricks is built on top of Apache Spark and provides a workspace for data engineering tasks.

    • It allows for easy integration with various data sources and tools for data processing.

    • Databricks provides features like notebooks, clusters, and libraries for efficient data engineering workflows.

  • Answered by AI

Skills evaluated in this interview

Interview experience
4
Good
Difficulty level
Moderate
Process Duration
2-4 weeks
Result
Not Selected

I applied via Naukri.com and was interviewed in Sep 2024. There was 1 interview round.

Round 1 - Coding Test 

Spark Optimization, Transformation, DLT, DL, Data Governance
Python
SQL

Interview Preparation Tips

Interview preparation tips for other job seekers - Ingestion, Integration, Spark, Optimization, Python, SQL, Data Warehouse
Interview experience
4
Good
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
-

I applied via LinkedIn and was interviewed in Feb 2024. There were 3 interview rounds.

Round 1 - Technical 

(2 Questions)

  • Q1. How to work with nested json using pyspark
  • Ans. 

    Working with nested JSON using PySpark involves using the StructType and StructField classes to define the schema and then using the select function to access nested fields.

    • Define the schema using StructType and StructField classes

    • Use the select function to access nested fields

    • Use dot notation to access nested fields, for example df.select('nested_field.sub_field')

  • Answered by AI
  • Q2. How to implement scd2 step by step
  • Ans. 

    Implementing SCD2 involves tracking historical changes in data over time.

    • Identify the business key that uniquely identifies each record

    • Add effective start and end dates to track when the record was valid

    • Insert new records with updated data and end date of '9999-12-31'

    • Update end date of previous record when a change occurs

  • Answered by AI
Round 2 - Technical 

(2 Questions)

  • Q1. Write a SQL query to select data from table 2 where data exists in table 1
  • Ans. 

    Use a SQL query to select data from table 2 where data exists in table 1

    • Use a JOIN statement to link the two tables based on a common column

    • Specify the columns you want to select from table 2

    • Use a WHERE clause to check for existence of data in table 1

  • Answered by AI
  • Q2. After performing joins how many records would be retrieved for inner, left, right and outer joins
  • Ans. 

    The number of records retrieved after performing joins depends on the type of join - inner, left, right, or outer.

    • Inner join retrieves only the matching records from both tables

    • Left join retrieves all records from the left table and matching records from the right table

    • Right join retrieves all records from the right table and matching records from the left table

    • Outer join retrieves all records from both tables, filling

  • Answered by AI
Round 3 - HR 

(1 Question)

  • Q1. About previous company and reason for leaving

Interview Preparation Tips

Interview preparation tips for other job seekers - Don't be afraid of giving interviews. Prepare well attend confidently if you clear it's an opportunity if you don't it's an experience!!

Skills evaluated in this interview

Interview experience
4
Good
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
No response

I applied via Recruitment Consulltant and was interviewed in Mar 2024. There were 2 interview rounds.

Round 1 - One-on-one 

(1 Question)

  • Q1. How the files will be read in AWS GLUE
  • Ans. 

    Files can be read in AWS Glue using Data Catalog, crawlers, and Glue ETL jobs.

    • Use AWS Glue Data Catalog to store metadata information about the files.

    • Crawlers can automatically infer the schema of the files and populate the Data Catalog.

    • Glue ETL jobs can then be used to read the files from various sources like S3, RDS, etc.

    • Supports various file formats like CSV, JSON, Parquet, etc.

  • Answered by AI
Round 2 - Technical 

(2 Questions)

  • Q1. How to get duplicate records
  • Ans. 

    Duplicate records can be identified using SQL queries by comparing columns and using aggregate functions.

    • Use GROUP BY clause with COUNT() function to identify duplicate records

    • Use HAVING clause to filter out records with count greater than 1

    • Join the table with itself on specific columns to find duplicates

  • Answered by AI
  • Q2. How to get the second highest salary

Skills evaluated in this interview

Interview experience
4
Good
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Technical 

(2 Questions)

  • Q1. Previous projects
  • Q2. Basic questions on python, azure cloud
Round 2 - HR 

(2 Questions)

  • Q1. Why company change
  • Q2. Salary negotiations
Interview experience
3
Average
Difficulty level
Easy
Process Duration
2-4 weeks
Result
Not Selected

I was interviewed in Apr 2024.

Round 1 - Technical 

(1 Question)

  • Q1. Tell me about yourself; Projects done; What is Columnar format file in Spark; Internals of Spark, Difference between OLAP and OLTP; About Datawarehouse- facts, dimensions
  • Ans. 

    I am a Senior Data Engineer with experience in various projects involving columnar format files in Spark, understanding Spark internals, OLAP vs OLTP, and data warehousing concepts.

    • Projects: Developed ETL pipelines using Spark for processing large datasets, implemented data quality checks, and optimized query performance.

    • Columnar format file in Spark: It stores data in columnar format to improve query performance by re...

  • Answered by AI

Skills evaluated in this interview

Infoqort Technologies Interview FAQs

How many rounds are there in Infoqort Technologies Senior Data Engineer interview?
Infoqort Technologies interview process usually has 3 rounds. The most common rounds in the Infoqort Technologies interview process are One-on-one Round and Resume Shortlist.
What are the top questions asked in Infoqort Technologies Senior Data Engineer interview?

Some of the top questions asked at the Infoqort Technologies Senior Data Engineer interview -

  1. Nothing njdfhjjbhjhg...read more
  2. Nothing hhhhhhhjjj...read more
  3. Fghjnnkkknvccccggb...read more

Tell us how to improve this page.

Infoqort Technologies Senior Data Engineer Reviews and Ratings

based on 1 review

1.0/5

Rating in categories

1.0

Skill development

1.0

Work-Life balance

1.0

Salary & Benefits

1.0

Job Security

1.0

Company culture

1.0

Promotions/Appraisal

1.0

Work Satisfaction

Explore 1 Review and Rating
Power BI Developer
14 salaries
unlock blur

₹3.6 L/yr - ₹10.4 L/yr

Data Engineer
6 salaries
unlock blur

₹3 L/yr - ₹5.4 L/yr

Sharepoint Developers
5 salaries
unlock blur

₹4.5 L/yr - ₹7.5 L/yr

Software Developer
4 salaries
unlock blur

₹1.8 L/yr - ₹4.5 L/yr

BI Developer
4 salaries
unlock blur

₹4.8 L/yr - ₹11.6 L/yr

Explore more salaries
Compare Infoqort Technologies with

TCS

3.7
Compare

Wipro

3.7
Compare

HCLTech

3.5
Compare

Tech Mahindra

3.6
Compare

Calculate your in-hand salary

Confused about how your in-hand salary is calculated? Enter your annual salary (CTC) and get your in-hand salary
Did you find this page helpful?
Yes No
write
Share an Interview