i
Cognizant
Proud winner of ABECA 2024 - AmbitionBox Employee Choice Awards
Filter interviews by
Aptitude test involved with quantative aptitude, logical reasoning and reading comprehensions.
I have strong skills in data processing, ETL, data modeling, and programming languages like Python and SQL.
Proficient in data processing and ETL techniques
Strong knowledge of data modeling and database design
Experience with programming languages like Python and SQL
Familiarity with big data technologies such as Hadoop and Spark
Yes, I am open to relocating for the right opportunity.
I am willing to relocate for the right job opportunity.
I have experience moving for previous roles.
I am flexible and adaptable to new locations.
I am excited about the possibility of exploring a new city or country.
Different types of joins in Spark include inner join, outer join, left join, right join, and full join.
Inner join: Returns only the rows that have matching values in both datasets.
Outer join: Returns all rows when there is a match in either dataset.
Left join: Returns all rows from the left dataset and the matched rows from the right dataset.
Right join: Returns all rows from the right dataset and the matched rows from t...
Optimization techniques in Spark improve performance and efficiency of data processing.
Partitioning data to distribute workload evenly
Caching frequently accessed data in memory
Using broadcast variables for small lookup tables
Avoiding shuffling operations whenever possible
Delta Lake is an open-source storage layer that brings ACID transactions to Apache Spark and big data workloads.
Delta Lake provides ACID transactions, scalable metadata handling, and unifies streaming and batch data processing.
It stores data in Parquet format and uses Apache Spark for processing.
Delta Lake ensures data reliability and data quality by providing schema enforcement and data versioning.
It supports time tra...
Tuning operations in Databricks involves optimizing performance and efficiency of data processing tasks.
Use cluster configuration settings to allocate resources efficiently
Optimize code by minimizing data shuffling and reducing unnecessary operations
Leverage Databricks Auto Optimize to automatically tune performance
Monitor job performance using Databricks Runtime Metrics and Spark UI
What people are saying about Cognizant
I applied via Approached by Company and was interviewed in Jun 2024. There was 1 interview round.
Spark optimization techniques used in project
Partitioning data to optimize parallel processing
Caching frequently accessed data to reduce computation time
Using broadcast variables for efficient data sharing across nodes
Optimizing shuffle operations to minimize data movement
Tuning memory and CPU settings for better performance
XCom in Airflow is a way for tasks to exchange messages or small amounts of data.
XCom allows tasks to communicate with each other by passing small pieces of data
It can be used to share information between tasks in a DAG
XCom can be used to pass information like task status, results, or any other data
To connect S3 from Databricks, you can use the AWS connector provided by Databricks.
Use the AWS connector provided by Databricks to connect to S3
Provide the necessary AWS credentials and S3 bucket details in the connector configuration
You can access S3 data using the file system API in Databricks
CDC stands for Change Data Capture, a process of identifying and capturing changes made to data in a database.
CDC is used to track changes in data over time, allowing for real-time data integration and analysis.
It captures inserts, updates, and deletes made to data, providing a historical record of changes.
CDC is commonly used in data warehousing, data replication, and data integration processes.
Examples of CDC tools i...
Cognizant interview questions for designations
I applied via Campus Placement
I want to join Cognizant because of its reputation for innovation and growth opportunities.
Cognizant is known for its cutting-edge technology solutions
I am impressed by Cognizant's commitment to employee development
I believe Cognizant will provide me with a challenging and rewarding work environment
I am passionate about working with data and enjoy the challenges and opportunities that come with being a data engineer.
I have a strong background in data engineering and enjoy working with data processing technologies such as Hadoop, Spark, and Kafka.
I find data engineering to be a dynamic and evolving field that allows me to continuously learn and grow my skills.
I am excited about the impact that data engineering can...
Get interview-ready with Top Cognizant Interview Questions
I applied via Naukri.com and was interviewed in Jan 2024. There was 1 interview round.
Word count by spark, flatMap, and map difference
Spark is a distributed computing framework for big data processing
flatMap is used to split each input string into words
map is used to transform each word into a key-value pair for counting
The difference lies in how the data is processed and transformed
Flat map is used to flatten nested arrays while map is used to transform each element in an array.
Flat map is used to flatten nested arrays into a single array.
Map is used to transform each element in an array using a function.
Flat map is commonly used in functional programming languages like JavaScript and Scala.
Map is a higher-order function that applies a given function to each element in an array.
Partitioning is dividing data into smaller chunks for better organization and performance, while bucketing is grouping data based on a specific criteria.
Partitioning is dividing data into smaller subsets based on a column or key.
Bucketing is grouping data based on a specific number of buckets or ranges.
Partitioning is commonly used in distributed systems for better data organization and query performance.
Bucketing is o...
50 MCQ for python SQL
I applied via Company Website and was interviewed in Oct 2023. There was 1 interview round.
Supervised learning uses labeled data to train the model, while unsupervised learning uses unlabeled data.
Supervised learning requires a target variable to be predicted, while unsupervised learning does not.
In supervised learning, the model learns from labeled training data, whereas in unsupervised learning, the model finds patterns in unlabeled data.
Examples of supervised learning include regression and classification...
Object Oriented Programming in Python focuses on creating classes and objects to organize code and data.
Python supports classes, objects, inheritance, polymorphism, and encapsulation.
Classes are blueprints for creating objects, which are instances of classes.
Inheritance allows a class to inherit attributes and methods from another class.
Polymorphism enables objects to be treated as instances of their parent class.
Encap...
To find delta between two tables in SQL, use the EXCEPT or MINUS operator.
Use the EXCEPT operator in SQL to return rows from the first table that do not exist in the second table.
Use the MINUS operator in SQL to return distinct rows from the first table that do not exist in the second table.
Exception handling in Python allows for graceful handling of errors and preventing program crashes.
Use try-except blocks to catch and handle exceptions.
Multiple except blocks can be used to handle different types of exceptions.
Finally block can be used to execute code regardless of whether an exception was raised or not.
Custom exceptions can be defined by creating a new class that inherits from the built-in Exception c
Decorators in Python are functions that modify the behavior of other functions.
Decorators are defined using the @decorator_name syntax before the function definition.
They can be used for logging, timing, authentication, etc.
Example: @staticmethod decorator in Python makes a method static.
Different types of joins in SQL include inner join, left join, right join, and full outer join.
Inner join: Returns rows when there is a match in both tables.
Left join: Returns all rows from the left table and the matched rows from the right table.
Right join: Returns all rows from the right table and the matched rows from the left table.
Full outer join: Returns rows when there is a match in either table.
Various optimizations such as indexing, caching, and parallel processing were used in the project.
Implemented indexing on frequently queried columns to improve query performance
Utilized caching mechanisms to store frequently accessed data and reduce database load
Implemented parallel processing to speed up data processing tasks
Optimized algorithms and data structures for efficient data retrieval and manipulation
I optimized code, increased memory allocation, used efficient data structures, and implemented data partitioning.
Optimized code by identifying and fixing memory leaks
Increased memory allocation for the application
Used efficient data structures like arrays, hashmaps, and trees
Implemented data partitioning to distribute data across multiple nodes
The duration of Cognizant Data Engineer interview process can vary, but typically it takes about less than 2 weeks to complete.
based on 32 interviews
2 Interview rounds
based on 139 reviews
Rating in categories
Associate
72.2k
salaries
| ₹5.1 L/yr - ₹16 L/yr |
Programmer Analyst
55.5k
salaries
| ₹2.4 L/yr - ₹9.7 L/yr |
Senior Associate
48.6k
salaries
| ₹9 L/yr - ₹27.3 L/yr |
Senior Processing Executive
28.9k
salaries
| ₹1.8 L/yr - ₹9 L/yr |
Technical Lead
17.6k
salaries
| ₹5.9 L/yr - ₹24.8 L/yr |
TCS
Infosys
Wipro
Accenture