Upload Button Icon Add office photos

Filter interviews by

ESG Book Data Intern Interview Questions and Answers

Updated 26 Jul 2024

ESG Book Data Intern Interview Experiences

1 interview found

Data Intern Interview Questions & Answers

user image Bhavani Shankar

posted on 26 Jul 2024

Interview experience
5
Excellent
Difficulty level
Easy
Process Duration
-
Result
Selected Selected
Round 1 - Coding Test 

Data analysis: Pandas, Excel

Interview Preparation Tips

Interview preparation tips for other job seekers - Best of the best people in england team. Very humble, polite and loving people.

Interview questions from similar companies

Data Intern Interview Questions & Answers

LTIMindtree user image mohammed rifat khan

posted on 31 Jul 2023

Interview experience
4
Good
Difficulty level
Easy
Process Duration
2-4 weeks
Result
Selected Selected

I applied via campus placement at Motilal Nehru Institute National Institute of Technology (NIT), Allahabad and was interviewed before Jul 2022. There were 3 interview rounds.

Round 1 - Resume Shortlist 
Pro Tip by AmbitionBox:
Keep your resume crisp and to the point. A recruiter looks at your resume for an average of 6 seconds, make sure to leave the best impression.
View all tips
Round 2 - Coding Test 

Two easy DS questions, arrays and vector stuff, I used c++

Round 3 - HR 

(2 Questions)

  • Q1. Nothing much, asked about locations and stuff like that, again pretty simple questioning, no technical
  • Q2. Nothing more was asked, don't know what to add

Interview Preparation Tips

Interview preparation tips for other job seekers - Good luck man, and wish me good luck too lol
Interview experience
5
Excellent
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
-

I applied via Recruitment Consulltant

Round 1 - Technical 

(5 Questions)

  • Q1. Explain ETL pipeline ecosystem in Azure Databricks?
  • Q2. Star vs Snowflake schema, when to use?
  • Q3. Find Salary higher than Average department salary
  • Q4. Implementation of SCD2 table
  • Q5. How incremental loading is done
Interview experience
3
Average
Difficulty level
Moderate
Process Duration
2-4 weeks
Result
Not Selected

I applied via Approached by Company and was interviewed in Dec 2024. There was 1 interview round.

Round 1 - Technical 

(2 Questions)

  • Q1. Tell me the high level overview of dataguard installation?
  • Q2. What are your daily tasks adn what things you handel in your team?
Interview experience
3
Average
Difficulty level
Easy
Process Duration
Less than 2 weeks
Result
Not Selected

I applied via Company Website and was interviewed in Nov 2024. There were 2 interview rounds.

Round 1 - Technical 

(2 Questions)

  • Q1. Sql constrainsts, star schema, dml dcl commands
  • Q2. About cureent project and responsibilities
Round 2 - Technical 

(2 Questions)

  • Q1. Current projects and resposibilities
  • Q2. Where vs having, reason for job change

Interview Preparation Tips

Interview preparation tips for other job seekers - 1. Technical - about you current project and responsibilities, basic SQL question-constraints, starschema, DML DCL command, one sql query write.
2. Technical with senior manager- about project ,where vs having , reason of job change
Interview experience
4
Good
Difficulty level
Easy
Process Duration
Less than 2 weeks
Result
No response

I applied via Naukri.com and was interviewed in Nov 2024. There was 1 interview round.

Round 1 - Technical 

(6 Questions)

  • Q1. Can you introduce yourself and describe your current project experience?
  • Ans. 

    I am a Senior Data Engineer with experience in building scalable data pipelines and optimizing data processing workflows.

    • Experience in designing and implementing ETL processes using tools like Apache Spark and Airflow

    • Proficient in working with large datasets and optimizing query performance

    • Strong background in data modeling and database design

    • Worked on projects involving real-time data processing and streaming analytic

  • Answered by AI
  • Q2. Decorators in python
  • Ans. 

    Decorators in Python are functions that modify the behavior of other functions or methods.

    • Decorators are defined using the @decorator_name syntax before a function definition.

    • They can be used to add functionality to existing functions without modifying their code.

    • Decorators can be used for logging, timing, authentication, and more.

    • Example: @staticmethod decorator in Python is used to define a static method in a class.

  • Answered by AI
  • Q3. What is the SQL query to group by employee ID in order to combine the first name and last name with a space?
  • Ans. 

    SQL query to group by employee ID and combine first name and last name with a space

    • Use the GROUP BY clause to group by employee ID

    • Use the CONCAT function to combine first name and last name with a space

    • Select employee ID, CONCAT(first_name, ' ', last_name) AS full_name

  • Answered by AI
  • Q4. What are constructors in Python?
  • Ans. 

    Constructors in Python are special methods used for initializing objects. They are called automatically when a new instance of a class is created.

    • Constructors are defined using the __init__() method in a class.

    • They are used to initialize instance variables of a class.

    • Example: class Person: def __init__(self, name, age): self.name = name self.age = age person1 = Person('Alice', 30)

  • Answered by AI
  • Q5. Indexing in sql
  • Ans. 

    Indexing in SQL is a technique used to improve the performance of queries by creating a data structure that allows for faster retrieval of data.

    • Indexes are created on columns in a database table to speed up the retrieval of rows that match a certain condition in a WHERE clause.

    • Indexes can be created using CREATE INDEX statement in SQL.

    • Types of indexes include clustered indexes, non-clustered indexes, unique indexes, an...

  • Answered by AI
  • Q6. Why spark works well with parquet files?
  • Ans. 

    Spark works well with Parquet files due to its columnar storage format, efficient compression, and ability to push down filters.

    • Parquet files are columnar storage format, which aligns well with Spark's processing model of working on columns rather than rows.

    • Parquet files support efficient compression, reducing storage space and improving read performance in Spark.

    • Spark can push down filters to Parquet files, allowing f...

  • Answered by AI

Skills evaluated in this interview

Data Engineer Interview Questions & Answers

Genpact user image Sashikanta Parida

posted on 17 Dec 2024

Interview experience
5
Excellent
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Not Selected

I applied via Recruitment Consulltant and was interviewed in Nov 2024. There were 2 interview rounds.

Round 1 - Technical 

(3 Questions)

  • Q1. What are different type of joins available in Databricks?
  • Ans. 

    Different types of joins available in Databricks include inner join, outer join, left join, right join, and cross join.

    • Inner join: Returns only the rows that have matching values in both tables.

    • Outer join: Returns all rows when there is a match in either table.

    • Left join: Returns all rows from the left table and the matched rows from the right table.

    • Right join: Returns all rows from the right table and the matched rows ...

  • Answered by AI
  • Q2. How do you make your data pipeline fault tolerant?
  • Ans. 

    Implementing fault tolerance in a data pipeline involves redundancy, monitoring, and error handling.

    • Use redundant components to ensure continuous data flow

    • Implement monitoring tools to detect failures and bottlenecks

    • Set up automated alerts for immediate response to issues

    • Design error handling mechanisms to gracefully handle failures

    • Use checkpoints and retries to ensure data integrity

  • Answered by AI
  • Q3. What is AutoLoader?
  • Ans. 

    AutoLoader is a feature in data engineering that automatically loads data from various sources into a data warehouse or database.

    • Automates the process of loading data from different sources

    • Reduces manual effort and human error

    • Can be scheduled to run at specific intervals

    • Examples: Apache Nifi, AWS Glue

  • Answered by AI
Round 2 - Technical 

(2 Questions)

  • Q1. How do you connect to different services in Azure?
  • Ans. 

    To connect to different services in Azure, you can use Azure SDKs, REST APIs, Azure Portal, Azure CLI, and Azure PowerShell.

    • Use Azure SDKs for programming languages like Python, Java, C#, etc.

    • Utilize REST APIs to interact with Azure services programmatically.

    • Access and manage services through the Azure Portal.

    • Leverage Azure CLI for command-line interface interactions.

    • Automate tasks using Azure PowerShell scripts.

  • Answered by AI
  • Q2. What are linked Services?
  • Ans. 

    Linked Services are connections to external data sources or destinations in Azure Data Factory.

    • Linked Services define the connection information needed to connect to external data sources or destinations.

    • They can be used in Data Factory pipelines to read from or write to external systems.

    • Examples of Linked Services include Azure Blob Storage, Azure SQL Database, and Amazon S3.

  • Answered by AI
Interview experience
4
Good
Difficulty level
Easy
Process Duration
-
Result
-

I applied via Recruitment Consulltant and was interviewed in Nov 2024. There was 1 interview round.

Round 1 - Technical 

(7 Questions)

  • Q1. Difference between bigtable and bigquery.
  • Ans. 

    Bigtable is a NoSQL database for real-time analytics, while BigQuery is a fully managed data warehouse for running SQL queries.

    • Bigtable is a NoSQL database designed for real-time analytics and high throughput, while BigQuery is a fully managed data warehouse for running SQL queries.

    • Bigtable is used for storing large amounts of semi-structured data, while BigQuery is used for analyzing structured data using SQL queries.

    • ...

  • Answered by AI
  • Q2. How to remove duplicate rows from bigquery? find the month of a given date in bigquery.
  • Ans. 

    To remove duplicate rows from BigQuery, use the DISTINCT keyword. To find the month of a given date, use the EXTRACT function.

    • To remove duplicate rows, use SELECT DISTINCT * FROM table_name;

    • To find the month of a given date, use SELECT EXTRACT(MONTH FROM date_column) AS month_name FROM table_name;

    • Make sure to replace 'table_name' and 'date_column' with the appropriate values in your query.

  • Answered by AI
  • Q3. What operator is used in composer to move data from gcs to bq
  • Ans. 

    The operator used in Composer to move data from GCS to BigQuery is the GCS to BigQuery operator.

    • The GCS to BigQuery operator is used in Apache Airflow, which is the underlying technology of Composer.

    • This operator allows you to transfer data from Google Cloud Storage (GCS) to BigQuery.

    • You can specify the source and destination parameters in the operator to define the data transfer process.

  • Answered by AI
  • Q4. Write a code for this - input = [1,2,3,4] output = [1,4,9,16]
  • Ans. 

    Code to square each element in the input array.

    • Iterate through the input array and square each element.

    • Store the squared values in a new array to get the desired output.

  • Answered by AI
  • Q5. Dataflow vs dataproc.
  • Ans. 

    Dataflow is a fully managed stream and batch processing service, while Dataproc is a managed Apache Spark and Hadoop service.

    • Dataflow is a serverless data processing service that automatically scales to handle your data, while Dataproc is a managed Spark and Hadoop service that requires you to provision and manage clusters.

    • Dataflow is designed for both batch and stream processing, allowing you to process data in real-t...

  • Answered by AI
  • Q6. Architecture of bq. Query optimization techniques in bigquery.
  • Ans. 

    BigQuery architecture includes storage, execution, and optimization components for efficient query processing.

    • BigQuery stores data in Capacitor storage system for fast access.

    • Query execution is distributed across multiple nodes for parallel processing.

    • Query optimization techniques include partitioning tables, clustering tables, and using query cache.

    • Using partitioned tables can help eliminate scanning unnecessary data.

    • ...

  • Answered by AI
  • Q7. RDD vs dataframe vs dataset in pyspark
  • Ans. 

    RDD vs dataframe vs dataset in PySpark

    • RDD (Resilient Distributed Dataset) is the basic abstraction in PySpark, representing a distributed collection of objects

    • Dataframe is a distributed collection of data organized into named columns, similar to a table in a relational database

    • Dataset is a distributed collection of data with the ability to use custom classes for type safety and user-defined functions

    • Dataframes and Data...

  • Answered by AI
Interview experience
5
Excellent
Difficulty level
Hard
Process Duration
Less than 2 weeks
Result
Not Selected

I applied via Naukri.com and was interviewed in Nov 2024. There was 1 interview round.

Round 1 - One-on-one 

(5 Questions)

  • Q1. What about your self?
  • Q2. Family background
  • Q3. Power BI test and advanced excel
  • Q4. Microsoft access test
  • Q5. Python test and One to one discussion with super boss
Interview experience
5
Excellent
Difficulty level
Easy
Process Duration
Less than 2 weeks
Result
Selected Selected

I applied via Recruitment Consulltant and was interviewed in Nov 2024. There were 2 interview rounds.

Round 1 - HR 

(2 Questions)

  • Q1. Can you provide an overview of your background, including your past experiences and daily activities, as well as the tools you use in your routine?
  • Ans. 

    I have a background in data analysis with experience in using tools like Python, SQL, and Tableau.

    • I have a degree in Statistics and have worked as a Data Analyst for 3 years.

    • My daily activities include cleaning and analyzing data, creating visualizations, and presenting insights to stakeholders.

    • I use Python for data manipulation and analysis, SQL for querying databases, and Tableau for creating interactive dashboards.

    • I...

  • Answered by AI
  • Q2. What are the concepts of advanced Excel and Power BI projects, and how are they utilized within a company or for clients?
  • Ans. 

    Advanced Excel and Power BI are tools used for data analysis and visualization in companies and for clients.

    • Advanced Excel allows for complex data manipulation, analysis, and visualization using features like pivot tables, macros, and VBA programming.

    • Power BI is a business analytics tool that provides interactive visualizations and business intelligence capabilities, connecting to various data sources.

    • These tools are u...

  • Answered by AI
Round 2 - One-on-one 

(2 Questions)

  • Q1. Can you explain your project experience related to Advanced Excel and Power BI?
  • Ans. 

    I have extensive experience in using Advanced Excel and Power BI for data analysis projects.

    • Created complex formulas and macros in Excel to automate data processing tasks

    • Designed interactive dashboards in Power BI to visualize and analyze data trends

    • Integrated data from multiple sources into Power BI for comprehensive analysis

    • Used Power Query and Power Pivot in Excel to manipulate and analyze large datasets

    • Provided dat...

  • Answered by AI
  • Q2. What are the concepts of credit and operations, particularly in relation to Know Your Customer (KYC) procedures and the privacy of client data?
  • Ans. 

    Credit and operations concepts in relation to KYC procedures and client data privacy.

    • Credit refers to the extension of money or resources to a client based on their financial history and ability to repay.

    • Operations involve the day-to-day processes and procedures within a financial institution to ensure smooth functioning.

    • KYC procedures are used to verify the identity of clients to prevent fraud and money laundering.

    • Pri...

  • Answered by AI

Interview Preparation Tips

Interview preparation tips for other job seekers - If your resume is shortlisted, then there is a higher chance that you will be selected.

ESG Book Interview FAQs

How many rounds are there in ESG Book Data Intern interview?
ESG Book interview process usually has 1 rounds. The most common rounds in the ESG Book interview process are Coding Test.

Tell us how to improve this page.

Interview Questions from Similar Companies

TCS Interview Questions
3.7
 • 10.2k Interviews
Accenture Interview Questions
3.9
 • 8k Interviews
Infosys Interview Questions
3.7
 • 7.5k Interviews
Wipro Interview Questions
3.7
 • 5.5k Interviews
Cognizant Interview Questions
3.8
 • 5.5k Interviews
Amazon Interview Questions
4.1
 • 4.9k Interviews
Capgemini Interview Questions
3.8
 • 4.7k Interviews
Tech Mahindra Interview Questions
3.6
 • 3.8k Interviews
HCLTech Interview Questions
3.5
 • 3.7k Interviews
Genpact Interview Questions
3.9
 • 3k Interviews
View all
Senior Analyst
13 salaries
unlock blur

₹5.5 L/yr - ₹7 L/yr

Senior Data Analyst
8 salaries
unlock blur

₹5.7 L/yr - ₹6.4 L/yr

Assistant Manager
6 salaries
unlock blur

₹9 L/yr - ₹14 L/yr

Analyst
5 salaries
unlock blur

₹5 L/yr - ₹6 L/yr

Data Analyst
5 salaries
unlock blur

₹4.8 L/yr - ₹5.2 L/yr

Explore more salaries
Compare ESG Book with

Sustainalytics

2.8
Compare

Institutional Shareholder Services

3.8
Compare

TCS

3.7
Compare

Accenture

3.9
Compare

Calculate your in-hand salary

Confused about how your in-hand salary is calculated? Enter your annual salary (CTC) and get your in-hand salary
Did you find this page helpful?
Yes No
write
Share an Interview