i
Filter interviews by
I applied via Job Portal and was interviewed in Feb 2021. There were 5 interview rounds.
To troubleshoot hive slowness, check for resource contention, optimize queries, and monitor system performance.
Check for resource contention such as CPU, memory, and disk usage
Optimize queries by reducing data scanned and avoiding unnecessary joins
Monitor system performance using tools like Ganglia or Ambari
Consider partitioning tables to improve query performance
Use compression to reduce disk I/O and network traffic
ZK and JN play a crucial role in ensuring high availability of Namenode in Hadoop cluster.
ZK (Zookeeper) is used for leader election and maintaining the state of active and standby Namenodes.
JN (JournalNode) is responsible for storing the edit logs of Namenode, which are used for recovery in case of failover.
In case of Namenode failure, ZK helps in electing a new active Namenode from the standby Namenodes, while JN pro...
Top trending discussions
I applied via Referral and was interviewed in Nov 2024. There were 2 interview rounds.
I applied via Campus Placement and was interviewed in Oct 2024. There were 2 interview rounds.
IT WAS Difficult and i gave up
I applied via Naukri.com and was interviewed in Aug 2024. There was 1 interview round.
The data analysis process for an employee salary table involves data cleaning, exploration, visualization, and modeling.
Start by cleaning the data to remove any duplicates, missing values, or outliers.
Explore the data by calculating basic statistics like mean, median, and standard deviation.
Visualize the data using histograms, box plots, or scatter plots to identify patterns or trends.
Perform statistical modeling to an...
I faced challenges in managing large datasets and ensuring data accuracy in my project.
Managing large volumes of data from multiple sources
Ensuring data accuracy and consistency
Dealing with missing or incomplete data
Handling data cleaning and preprocessing tasks
Addressing technical issues with data storage or processing tools
posted on 8 Nov 2024
To calculate TDI and MDI of a customer in SQL, you need to aggregate and analyze the customer's transaction data.
Calculate Total Daily Income (TDI) by summing up the income of the customer for each day.
Calculate Monthly Daily Income (MDI) by averaging the TDI for each day in a month.
Use SQL functions like SUM, AVG, and GROUP BY to perform the calculations.
Example: SELECT SUM(income) AS TDI, AVG(income) AS MDI FROM tran
They asked about the sas,sql,and power bi
I applied via Naukri.com and was interviewed in Sep 2024. There were 4 interview rounds.
Basic aptitude questions
Data structure and algorithms
I applied via campus placement at Indian Institute of Technology (IIT), Jodhpur and was interviewed in Jun 2024. There was 1 interview round.
A p-value is a measure used in statistical hypothesis testing to determine the strength of evidence against the null hypothesis.
A p-value is the probability of obtaining results as extreme as the observed results, assuming the null hypothesis is true.
A p-value is compared to a significance level (usually 0.05) to determine if the null hypothesis should be rejected.
A p-value less than the significance level indicates st
The output of a**2 is the square of the value of a.
The output is the value of a multiplied by itself
For example, if a = 3, then the output would be 9 (3*3)
append() adds elements to a single DataFrame, while concat() combines multiple DataFrames.
append() is a method used to add rows to a DataFrame.
concat() is a function used to combine multiple DataFrames along a particular axis.
append() modifies the original DataFrame, while concat() returns a new DataFrame.
Example: df1.append(df2) vs pd.concat([df1, df2])
I applied via Referral and was interviewed in Oct 2024. There was 1 interview round.
Basic aptitude test like distance problem , age etc
I applied via Naukri.com and was interviewed in Mar 2024. There were 3 interview rounds.
Error handling in PySpark involves using try-except blocks and logging to handle exceptions and errors.
Use try-except blocks to catch and handle exceptions in PySpark code
Utilize logging to record errors and exceptions for debugging purposes
Consider using the .option('mode', 'PERMISSIVE') method to handle corrupt records in data processing
Senior Analyst
250
salaries
| ₹8.5 L/yr - ₹36.5 L/yr |
Senior Associate
216
salaries
| ₹5.8 L/yr - ₹12 L/yr |
Senior Software Engineer
193
salaries
| ₹15 L/yr - ₹50.1 L/yr |
Analyst
191
salaries
| ₹5 L/yr - ₹20 L/yr |
Software Engineer
175
salaries
| ₹11 L/yr - ₹35.6 L/yr |
HDFC Bank
ICICI Bank
State Bank of India
Axis Bank