Upload Button Icon Add office photos
Engaged Employer

i

This company page is being actively managed by CGI Group Team. If you also belong to the team, you can get access from here

CGI Group Verified Tick

Compare button icon Compare button icon Compare

Proud winner of ABECA 2024 - AmbitionBox Employee Choice Awards

zig zag pattern zig zag pattern

Filter interviews by

CGI Group Senior Data Engineer Interview Questions, Process, and Tips

Updated 18 Dec 2024

Top CGI Group Senior Data Engineer Interview Questions and Answers

  • Q1. How do you utilize the enhanced optimization option in AWS Glue?
  • Q2. What are the best practices for optimizing querying in Amazon Redshift?
  • Q3. How to identify the root-cause if an executor is taking longer time compared to others

CGI Group Senior Data Engineer Interview Experiences

2 interviews found

Interview experience
3
Average
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Not Selected

I applied via Naukri.com and was interviewed in Nov 2024. There was 1 interview round.

Round 1 - Technical 

(2 Questions)

  • Q1. How do you utilize the enhanced optimization option in AWS Glue?
  • Ans. 

    Enhanced optimization in AWS Glue improves job performance by automatically adjusting resources based on workload

    • Enhanced optimization in AWS Glue automatically adjusts resources like DPUs based on workload

    • It helps improve job performance by optimizing resource allocation

    • Users can enable enhanced optimization in AWS Glue job settings

  • Answered by AI
  • Q2. What are the best practices for optimizing querying in Amazon Redshift?
  • Ans. 

    Optimizing querying in Amazon Redshift involves proper table design, distribution keys, sort keys, and query optimization techniques.

    • Use appropriate distribution keys to evenly distribute data across nodes for parallel processing.

    • Utilize sort keys to physically order data on disk, reducing the need for sorting during queries.

    • Avoid using SELECT * and instead specify only the columns needed to reduce data transfer.

    • Use AN...

  • Answered by AI
Interview experience
4
Good
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Technical 

(1 Question)

  • Q1. SQL Order of execution
  • Ans. 

    SQL statements are executed in a specific order to ensure accurate results.

    • SQL statements are executed in the following order: FROM, WHERE, GROUP BY, HAVING, SELECT, ORDER BY.

    • The FROM clause specifies the tables involved in the query.

    • The WHERE clause filters the rows based on specified conditions.

    • The GROUP BY clause groups the rows based on specified columns.

    • The HAVING clause filters the groups based on specified condi...

  • Answered by AI
Round 2 - Technical 

(1 Question)

  • Q1. How to identify the root-cause if an executor is taking longer time compared to others
  • Ans. 

    Identifying root cause of slow executor compared to others

    • Check resource utilization of the slow executor (CPU, memory, disk)

    • Look for any specific tasks or stages that are taking longer on the slow executor

    • Check for network latency or communication issues affecting the slow executor

    • Monitor garbage collection and JVM metrics for potential bottlenecks

    • Consider data skew or unbalanced data distribution causing slow perform

  • Answered by AI

Skills evaluated in this interview

Senior Data Engineer Interview Questions Asked at Other Companies

asked in 7 Eleven
Q1. Write a query to get the customer with the highest total order va ... read more
asked in 7 Eleven
Q2. There are 10 million records in the table and the schema does not ... read more
asked in 7 Eleven
Q3. How do you handle data pipeline when the schema information keeps ... read more
asked in 7 Eleven
Q4. Difference between Parquet and ORC file. Why industry uses parque ... read more
asked in 7 Eleven
Q5. What is Normalisation and Denormalisation? When do we use them? G ... read more

Interview questions from similar companies

Interview experience
5
Excellent
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Selected Selected

I applied via Naukri.com and was interviewed in Nov 2024. There were 2 interview rounds.

Round 1 - Technical 

(2 Questions)

  • Q1. All questions based on databricks and some pyspark , python , SQL only.
  • Q2. Learn windows function implementation in databricks note book.
Round 2 - HR 

(1 Question)

  • Q1. This round about your salary discussion around.

Interview Preparation Tips

Topics to prepare for Accenture Senior Data Engineer interview:
  • Python
  • Pyspark
  • SQL
  • Databricks
Interview preparation tips for other job seekers - Please prepare for pyspark, python , SQL , databricks for practice to switch your job to big data engineer
Interview experience
4
Good
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
No response

I was interviewed in Sep 2024.

Round 1 - Technical 

(3 Questions)

  • Q1. Coding in pyspark
  • Ans. 

    Pyspark is a Python library for big data processing using Spark framework.

    • Pyspark is used for processing large datasets in parallel.

    • It provides APIs for data manipulation, querying, and analysis.

    • Example: Using pyspark to read a CSV file and perform data transformations.

  • Answered by AI
  • Q2. Databricks optimisation technique
  • Ans. 

    Databricks optimisation techniques improve performance and efficiency of data processing on the Databricks platform.

    • Use cluster sizing and autoscaling to optimize resource allocation based on workload

    • Leverage Databricks Delta for optimized data storage and processing

    • Utilize caching and persisting data to reduce computation time

    • Optimize queries by using appropriate indexing and partitioning strategies

  • Answered by AI
  • Q3. Aqe details in databricks
  • Ans. 

    Databricks is a unified data analytics platform that provides a collaborative environment for data engineers.

    • Databricks is built on top of Apache Spark and provides a workspace for data engineering tasks.

    • It allows for easy integration with various data sources and tools for data processing.

    • Databricks provides features like notebooks, clusters, and libraries for efficient data engineering workflows.

  • Answered by AI

Skills evaluated in this interview

Interview experience
4
Good
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Technical 

(1 Question)

  • Q1. Adf,etl,python,adb
Interview experience
5
Excellent
Difficulty level
Hard
Process Duration
2-4 weeks
Result
-

I applied via Naukri.com and was interviewed in Sep 2024. There was 1 interview round.

Round 1 - One-on-one 

(2 Questions)

  • Q1. What is scd type 2?
  • Ans. 

    SCD type 2 is a method used in data warehousing to track historical changes by creating a new record for each change.

    • SCD type 2 stands for Slowly Changing Dimension type 2

    • It involves creating a new record in the dimension table whenever there is a change in the data

    • The old record is marked as inactive and the new record is marked as current

    • It allows for historical tracking of changes in data over time

    • Example: If a cust...

  • Answered by AI
  • Q2. Pyspark question read CSV from folder and add column in each csv file and write it to different location.

Skills evaluated in this interview

Interview experience
3
Average
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Technical 

(4 Questions)

  • Q1. Explain about yourself.
  • Ans. 

    I am a Senior Data Engineer with 5+ years of experience in designing and implementing data pipelines for large-scale projects.

    • Experienced in ETL processes and data warehousing

    • Proficient in programming languages like Python, SQL, and Java

    • Skilled in working with big data technologies such as Hadoop, Spark, and Kafka

    • Strong understanding of data modeling and database management

    • Excellent problem-solving and communication sk

  • Answered by AI
  • Q2. Explain about your current project.
  • Ans. 

    Developing a real-time data processing system for analyzing customer behavior on e-commerce platform.

    • Utilizing Apache Kafka for real-time data streaming

    • Implementing Spark for data processing and analysis

    • Creating machine learning models for customer segmentation

    • Integrating with Elasticsearch for data indexing and search functionality

  • Answered by AI
  • Q3. Big data concept
  • Q4. Spark related questions

Interview Preparation Tips

Interview preparation tips for other job seekers - Explore more technical and project process.
Interview experience
4
Good
Difficulty level
Moderate
Process Duration
2-4 weeks
Result
Not Selected

I applied via Naukri.com and was interviewed in Sep 2024. There was 1 interview round.

Round 1 - Coding Test 

Spark Optimization, Transformation, DLT, DL, Data Governance
Python
SQL

Interview Preparation Tips

Interview preparation tips for other job seekers - Ingestion, Integration, Spark, Optimization, Python, SQL, Data Warehouse
Interview experience
3
Average
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Not Selected

I applied via Job Portal and was interviewed in Jul 2024. There was 1 interview round.

Round 1 - HR 

(2 Questions)

  • Q1. Tell about experience
  • Ans. 

    I have over 5 years of experience in data engineering, working with large datasets and implementing data pipelines.

    • Developed and maintained ETL processes to extract, transform, and load data from various sources

    • Optimized database performance and implemented data quality checks

    • Worked with cross-functional teams to design and implement data solutions

    • Utilized tools such as Apache Spark, Hadoop, and SQL for data processing

    • ...

  • Answered by AI
  • Q2. What would you do when you are given a new task to do
  • Ans. 

    I would start by understanding the requirements, breaking down the task into smaller steps, researching if needed, and then creating a plan to execute the task efficiently.

    • Understand the requirements of the task

    • Break down the task into smaller steps

    • Research if needed to gather necessary information

    • Create a plan to execute the task efficiently

    • Communicate with stakeholders for clarification or updates

    • Regularly track prog

  • Answered by AI
Interview experience
4
Good
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
-

I applied via LinkedIn and was interviewed in Feb 2024. There were 3 interview rounds.

Round 1 - Technical 

(2 Questions)

  • Q1. How to work with nested json using pyspark
  • Ans. 

    Working with nested JSON using PySpark involves using the StructType and StructField classes to define the schema and then using the select function to access nested fields.

    • Define the schema using StructType and StructField classes

    • Use the select function to access nested fields

    • Use dot notation to access nested fields, for example df.select('nested_field.sub_field')

  • Answered by AI
  • Q2. How to implement scd2 step by step
  • Ans. 

    Implementing SCD2 involves tracking historical changes in data over time.

    • Identify the business key that uniquely identifies each record

    • Add effective start and end dates to track when the record was valid

    • Insert new records with updated data and end date of '9999-12-31'

    • Update end date of previous record when a change occurs

  • Answered by AI
Round 2 - Technical 

(2 Questions)

  • Q1. Write a SQL query to select data from table 2 where data exists in table 1
  • Ans. 

    Use a SQL query to select data from table 2 where data exists in table 1

    • Use a JOIN statement to link the two tables based on a common column

    • Specify the columns you want to select from table 2

    • Use a WHERE clause to check for existence of data in table 1

  • Answered by AI
  • Q2. After performing joins how many records would be retrieved for inner, left, right and outer joins
  • Ans. 

    The number of records retrieved after performing joins depends on the type of join - inner, left, right, or outer.

    • Inner join retrieves only the matching records from both tables

    • Left join retrieves all records from the left table and matching records from the right table

    • Right join retrieves all records from the right table and matching records from the left table

    • Outer join retrieves all records from both tables, filling

  • Answered by AI
Round 3 - HR 

(1 Question)

  • Q1. About previous company and reason for leaving

Interview Preparation Tips

Interview preparation tips for other job seekers - Don't be afraid of giving interviews. Prepare well attend confidently if you clear it's an opportunity if you don't it's an experience!!

Skills evaluated in this interview

CGI Group Interview FAQs

How many rounds are there in CGI Group Senior Data Engineer interview?
CGI Group interview process usually has 1-2 rounds. The most common rounds in the CGI Group interview process are Technical.
How to prepare for CGI Group Senior Data Engineer interview?
Go through your CV in detail and study all the technologies mentioned in your CV. Prepare at least two technologies or languages in depth if you are appearing for a technical interview at CGI Group. The most common topics and skills that interviewers at CGI Group expect are Python, Analytics, CGI, Data Modeling and Data Warehousing.
What are the top questions asked in CGI Group Senior Data Engineer interview?

Some of the top questions asked at the CGI Group Senior Data Engineer interview -

  1. How do you utilize the enhanced optimization option in AWS Gl...read more
  2. What are the best practices for optimizing querying in Amazon Redshi...read more
  3. How to identify the root-cause if an executor is taking longer time compared to...read more

Tell us how to improve this page.

CGI Group Senior Data Engineer Interview Process

based on 2 interviews

Interview experience

3.5
  
Good
View more
CGI Group Senior Data Engineer Salary
based on 22 salaries
₹8 L/yr - ₹26.6 L/yr
17% less than the average Senior Data Engineer Salary in India
View more details

CGI Group Senior Data Engineer Reviews and Ratings

based on 2 reviews

4.1/5

Rating in categories

3.0

Skill development

4.4

Work-life balance

3.7

Salary

3.9

Job security

4.0

Company culture

1.0

Promotions

3.1

Work satisfaction

Explore 2 Reviews and Ratings
Software Engineer
8.1k salaries
unlock blur

₹3.8 L/yr - ₹12 L/yr

Senior Software Engineer
7.2k salaries
unlock blur

₹6.1 L/yr - ₹22.5 L/yr

Lead Analyst
3.1k salaries
unlock blur

₹9.9 L/yr - ₹29.5 L/yr

Associate Software Engineer
1.9k salaries
unlock blur

₹2.5 L/yr - ₹6.6 L/yr

Senior Test Engineer
1.2k salaries
unlock blur

₹6 L/yr - ₹18.4 L/yr

Explore more salaries
Compare CGI Group with

TCS

3.7
Compare

Infosys

3.6
Compare

Wipro

3.7
Compare

HCLTech

3.5
Compare
Did you find this page helpful?
Yes No
write
Share an Interview