i
Capgemini
Proud winner of ABECA 2024 - AmbitionBox Employee Choice Awards
Filter interviews by
I applied via Campus Placement and was interviewed before Aug 2022. There were 3 interview rounds.
DISTANCE,ENGLISH GRAMMAR,TIME
What people are saying about Capgemini
Capgemini interview questions for designations
I was interviewed in Jan 2025.
Simple aptitude test
Medium python code and that was easy to understand
Circuit breaker design pattern is used to prevent a system from repeatedly trying to execute an operation that is likely to fail.
Circuit breaker monitors the number of failures and trips when a threshold is reached.
It can be reset after a certain period of time to check if the operation is successful again.
Helps in improving system stability and resilience by handling failures gracefully.
Transformer stage is a processing stage in IBM InfoSphere DataStage used for data transformation.
Used for transforming data from source to target in DataStage
Can perform various operations like filtering, aggregating, joining, etc.
Supports parallel processing for efficient data transformation
Coding test followed by technical discussion
Governer limits are restrictions set by Salesforce to prevent code from consuming too much resources.
Governer limits are restrictions set by Salesforce to ensure efficient use of resources.
They include limits on CPU time, heap size, SOQL queries, DML statements, etc.
Exceeding these limits can result in exceptions like 'Apex CPU time limit exceeded'.
Reverse a given string by iterating through it backwards and appending each character to a new string.
Create a new empty string to store the reversed string
Iterate through the original string from the last character to the first character
Append each character to the new string
Return the new string as the reversed string
Azure Databricks is a unified analytics platform that combines big data processing and machine learning.
Collaborative environment for data scientists, data engineers, and business analysts
Integrated with Azure services for data storage, processing, and analytics
Supports popular programming languages like Python, Scala, and SQL
Provides tools for data visualization and machine learning model development
Cache in Databricks is a mechanism to store intermediate results of computations for faster access.
Cache in Databricks is used to store intermediate results of computations in memory for faster access.
It helps in reducing the time taken to recompute the same data by storing it in memory.
Data can be cached at different levels such as DataFrame, RDD, or table.
Example: df.cache() will cache the DataFrame 'df' in memory fo
based on 4 interviews
Interview experience
based on 32 reviews
Rating in categories
Consultant
55.1k
salaries
| ₹5.1 L/yr - ₹17.5 L/yr |
Associate Consultant
50.8k
salaries
| ₹3 L/yr - ₹11.8 L/yr |
Senior Consultant
46.1k
salaries
| ₹7.5 L/yr - ₹25 L/yr |
Senior Analyst
20.8k
salaries
| ₹2 L/yr - ₹9 L/yr |
Senior Software Engineer
20.1k
salaries
| ₹3.5 L/yr - ₹12.5 L/yr |
Wipro
Accenture
Cognizant
TCS