Filter interviews by
I applied via LinkedIn and was interviewed before Feb 2023. There were 2 interview rounds.
I applied via Naukri.com and was interviewed in Oct 2024. There were 2 interview rounds.
Spark performance problems can arise due to inefficient code, data skew, resource constraints, and improper configuration.
Inefficient code can lead to slow performance, such as using collect() on large datasets.
Data skew can cause uneven distribution of data across partitions, impacting processing time.
Resource constraints like insufficient memory or CPU can result in slow Spark jobs.
Improper configuration settings, su...
I applied via Job Fair and was interviewed in Nov 2024. There were 2 interview rounds.
I applied via Naukri.com and was interviewed in Nov 2024. There was 1 interview round.
Enhanced optimization in AWS Glue improves job performance by automatically adjusting resources based on workload
Enhanced optimization in AWS Glue automatically adjusts resources like DPUs based on workload
It helps improve job performance by optimizing resource allocation
Users can enable enhanced optimization in AWS Glue job settings
Optimizing querying in Amazon Redshift involves proper table design, distribution keys, sort keys, and query optimization techniques.
Use appropriate distribution keys to evenly distribute data across nodes for parallel processing.
Utilize sort keys to physically order data on disk, reducing the need for sorting during queries.
Avoid using SELECT * and instead specify only the columns needed to reduce data transfer.
Use AN...
posted on 11 Dec 2024
PySpark is a Python API for Apache Spark, used for big data processing and analytics.
PySpark is a Python API for Apache Spark, a fast and general-purpose cluster computing system.
It allows for easy integration with Python libraries and provides high-level APIs in Python.
PySpark can be used for processing large datasets, machine learning, real-time data streaming, and more.
It supports various data sources such as HDFS, ...
PySpark is a Python API for Apache Spark, while Python is a general-purpose programming language.
PySpark is specifically designed for big data processing using Spark, while Python is a versatile programming language used for various applications.
PySpark allows for distributed computing and parallel processing, while Python is primarily used for sequential programming.
PySpark provides libraries and tools for working wit...
posted on 28 Sep 2024
I applied via Campus Placement and was interviewed in Aug 2024. There were 8 interview rounds.
Database Management system SQL and PlSQL
Database Base Management system SQL and PlSQL
Database Management system
Database Management system
Database Management system
Database Management system
Database Base Management system
posted on 22 Oct 2024
Primary and standby databases can be synced using Data Guard in Oracle.
Configure Data Guard on primary and standby databases
Enable redo transport services
Monitor and resolve any synchronization issues
Perform switchover or failover when necessary
Seeking new challenges and growth opportunities in a dynamic environment.
Looking for new challenges to enhance my skills and knowledge
Seeking growth opportunities for career advancement
Interested in working in a more dynamic and innovative environment
I applied via Naukri.com and was interviewed in Sep 2024. There was 1 interview round.
SCD type 2 is a method used in data warehousing to track historical changes by creating a new record for each change.
SCD type 2 stands for Slowly Changing Dimension type 2
It involves creating a new record in the dimension table whenever there is a change in the data
The old record is marked as inactive and the new record is marked as current
It allows for historical tracking of changes in data over time
Example: If a cust...
posted on 23 Jul 2024
To patch specific db components to reduce downtime, follow these steps:
Identify the specific components that need to be patched
Plan the patching process during a scheduled maintenance window
Backup the database before applying patches
Apply patches to the specific components one by one
Test the patched components to ensure they are functioning correctly
Monitor the database for any issues post-patching
I applied via Company Website and was interviewed in Jul 2024. There was 1 interview round.
Pods are the smallest deployable units in Kubernetes, consisting of one or more containers.
Pods are used to run and manage containers in Kubernetes
Each pod has its own unique IP address within the Kubernetes cluster
Pods can contain multiple containers that share resources and are scheduled together
Pods are ephemeral and can be easily created, destroyed, or replicated
Pods can be managed and scaled using Kubernetes contr
Front end Developer
8
salaries
| ₹7 L/yr - ₹9.2 L/yr |
Software Engineer
5
salaries
| ₹8 L/yr - ₹13.6 L/yr |
Quality Analyst
5
salaries
| ₹3.3 L/yr - ₹5 L/yr |
Data Annotation Engineer
5
salaries
| ₹3 L/yr - ₹4 L/yr |
Product Manager
4
salaries
| ₹17.5 L/yr - ₹17.5 L/yr |
Fractal Analytics
Mu Sigma
Algonomy
Tiger Analytics