Filter interviews by
A string can be reversed without affecting memory size by swapping characters from both ends.
Iterate through half of the string length
Swap the characters at the corresponding positions from both ends
Hypothesis testing is a statistical method to determine if there is enough evidence to support or reject a claim.
Hypothesis testing involves formulating a null hypothesis and an alternative hypothesis.
The null hypothesis assumes that there is no significant difference or relationship between variables.
The alternative hypothesis suggests that there is a significant difference or relationship between variables.
Distr...
Gradient boosting is a machine learning technique that combines multiple weak models to create a strong predictive model.
Gradient boosting is an ensemble method that iteratively adds new models to correct the errors made by previous models.
It is a type of boosting algorithm that focuses on reducing the residual errors in predictions.
Gradient boosting uses a loss function and gradient descent to optimize the model'...
XGBoost and AdaBoost are both boosting algorithms, but XGBoost is an optimized version of AdaBoost.
XGBoost is an optimized version of AdaBoost that uses gradient boosting.
AdaBoost combines weak learners into a strong learner by adjusting weights.
XGBoost uses a more advanced regularization technique called 'gradient boosting'.
XGBoost is known for its speed and performance in large-scale machine learning tasks.
Both ...
Addressing skewed training data in data science
Analyze the extent of skewness in the data
Consider resampling techniques like oversampling or undersampling
Apply appropriate evaluation metrics that are robust to class imbalance
Explore ensemble methods like bagging or boosting
Use synthetic data generation techniques like SMOTE
Consider feature engineering to improve model performance
Regularize the model to avoid overf...
The cost function for linear regression is mean squared error (MSE) and for logistic regression is log loss.
The cost function for linear regression is calculated by taking the average of the squared differences between the predicted and actual values.
The cost function for logistic regression is calculated using the logarithm of the predicted probabilities.
The goal of the cost function is to minimize the error betw...
Regularization is a technique used in machine learning to prevent overfitting by adding a penalty term to the loss function.
Regularization helps to reduce the complexity of a model by discouraging large parameter values.
It prevents overfitting by adding a penalty for complex models, encouraging simpler and more generalizable models.
Common regularization techniques include L1 regularization (Lasso), L2 regularizati...
The objective of predictive modeling is to minimize the cost function as it helps in optimizing the model's performance.
Predictive modeling aims to make accurate predictions by minimizing the cost function.
The cost function quantifies the discrepancy between predicted and actual values.
By minimizing the cost function, the model can improve its ability to make accurate predictions.
The cost function can be defined d...
Principal Component Analysis (PCA) is a dimensionality reduction technique used to transform high-dimensional data into a lower-dimensional space.
PCA is used to identify patterns and relationships in data by reducing the number of variables.
It helps in visualizing and interpreting complex data by representing it in a simpler form.
PCA is commonly used in fields like image processing, genetics, finance, and social s...
posted on 4 Dec 2016
I applied via Campus Placement and was interviewed in Jan 2016. There were 5 interview rounds.
I have 2 years of experience in data analytics, including working with large datasets and creating data visualizations.
Worked with large datasets to extract meaningful insights
Created data visualizations using tools like Tableau and Power BI
Utilized statistical analysis techniques to identify trends and patterns
Collaborated with cross-functional teams to drive data-driven decision making
I applied via Campus Placement and was interviewed in Dec 2016. There were 6 interview rounds.
I have a strong background in data analysis and machine learning with experience in various industries.
Bachelor's degree in Statistics with a focus on machine learning
Worked as a data analyst at XYZ company, where I developed predictive models to optimize marketing strategies
Internship at ABC company, where I analyzed customer data to improve retention rates
Proficient in programming languages such as Python and R
Analytics helps uncover insights from data to drive informed decision-making and improve business outcomes.
Analytics allows for data-driven decision-making
Helps identify trends and patterns in data
Enables businesses to optimize processes and strategies
Can lead to improved efficiency and effectiveness
Allows for predictive modeling and forecasting
Examples: using customer data to personalize marketing campaigns, analyzing...
Hypothesis testing is a statistical method to determine if there is enough evidence to support or reject a claim.
Hypothesis testing involves formulating a null hypothesis and an alternative hypothesis.
The null hypothesis assumes that there is no significant difference or relationship between variables.
The alternative hypothesis suggests that there is a significant difference or relationship between variables.
Distributi...
A string can be reversed without affecting memory size by swapping characters from both ends.
Iterate through half of the string length
Swap the characters at the corresponding positions from both ends
Gradient boosting is a machine learning technique that combines multiple weak models to create a strong predictive model.
Gradient boosting is an ensemble method that iteratively adds new models to correct the errors made by previous models.
It is a type of boosting algorithm that focuses on reducing the residual errors in predictions.
Gradient boosting uses a loss function and gradient descent to optimize the model's per...
XGBoost and AdaBoost are both boosting algorithms, but XGBoost is an optimized version of AdaBoost.
XGBoost is an optimized version of AdaBoost that uses gradient boosting.
AdaBoost combines weak learners into a strong learner by adjusting weights.
XGBoost uses a more advanced regularization technique called 'gradient boosting'.
XGBoost is known for its speed and performance in large-scale machine learning tasks.
Both algor...
Developed a machine learning model to predict customer churn for a telecom company
Collected and cleaned customer data including usage patterns and demographics
Used classification algorithms like Random Forest and Logistic Regression to build the model
Evaluated model performance using metrics like accuracy, precision, and recall
Implemented the model in a production environment for real-time predictions
Addressing skewed training data in data science
Analyze the extent of skewness in the data
Consider resampling techniques like oversampling or undersampling
Apply appropriate evaluation metrics that are robust to class imbalance
Explore ensemble methods like bagging or boosting
Use synthetic data generation techniques like SMOTE
Consider feature engineering to improve model performance
Regularize the model to avoid overfittin...
Principal Component Analysis (PCA) is a dimensionality reduction technique used to transform high-dimensional data into a lower-dimensional space.
PCA is used to identify patterns and relationships in data by reducing the number of variables.
It helps in visualizing and interpreting complex data by representing it in a simpler form.
PCA is commonly used in fields like image processing, genetics, finance, and social scienc...
The cost function for linear regression is mean squared error (MSE) and for logistic regression is log loss.
The cost function for linear regression is calculated by taking the average of the squared differences between the predicted and actual values.
The cost function for logistic regression is calculated using the logarithm of the predicted probabilities.
The goal of the cost function is to minimize the error between t...
Regularization is a technique used in machine learning to prevent overfitting by adding a penalty term to the loss function.
Regularization helps to reduce the complexity of a model by discouraging large parameter values.
It prevents overfitting by adding a penalty for complex models, encouraging simpler and more generalizable models.
Common regularization techniques include L1 regularization (Lasso), L2 regularization (R...
The objective of predictive modeling is to minimize the cost function as it helps in optimizing the model's performance.
Predictive modeling aims to make accurate predictions by minimizing the cost function.
The cost function quantifies the discrepancy between predicted and actual values.
By minimizing the cost function, the model can improve its ability to make accurate predictions.
The cost function can be defined differ...
I chose your company because of its strong reputation and the opportunity to work on diverse projects.
Your company has a strong reputation in the industry.
I am impressed by the diverse range of projects your company is involved in.
Your company offers a collaborative and innovative work environment.
I believe working at your company will provide me with valuable hands-on experience.
Your company's commitment to profession...
Top trending discussions
I applied via Naukri.com and was interviewed in Oct 2020. There was 1 interview round.
posted on 8 Feb 2022
posted on 23 May 2024
I applied via Approached by Company and was interviewed in Apr 2024. There were 3 interview rounds.
Business process and overview
I appeared for an interview before Apr 2024, where I was asked the following questions.
I applied via LinkedIn and was interviewed in Sep 2023. There were 2 interview rounds.
Blend combines data from different data sources in a single view, while join combines data from the same data source.
Blend is used when data comes from different data sources, join is used when data comes from the same data source
Blend creates a relationship between data sources without actually merging the data, join merges the data based on a common field
Blend can be used to combine data from different databases like...
I applied via Job Portal and was interviewed before Mar 2023. There were 4 interview rounds.
I applied via Naukri.com and was interviewed before Feb 2022. There were 4 interview rounds.
Basic aptitude questions of quant and reasoning and general english
Case study of varies products asked around 15-20 which were easy to attempt
SQL is a programming language used for managing and manipulating relational databases. Excel formulas are used for performing calculations and data analysis in Microsoft Excel.
SQL is used to retrieve, insert, update, and delete data from databases.
Excel formulas are written using functions and operators to perform calculations and manipulate data.
SQL example: SELECT * FROM customers WHERE age > 30;
Excel formula example...
Some of the top questions asked at the Noodle Analytics Associate Data Scientist interview -
Data Scientist
10
salaries
| ₹10 L/yr - ₹21 L/yr |
Data Engineer
6
salaries
| ₹7 L/yr - ₹20 L/yr |
Cloud Engineer
6
salaries
| ₹9.5 L/yr - ₹14 L/yr |
Senior Data Scientist
6
salaries
| ₹18 L/yr - ₹28 L/yr |
Software Engineer
5
salaries
| ₹9.6 L/yr - ₹25 L/yr |
Rapid Care Transcription
Carisma Solutions
GfK MODE
Edward Food Research and Analysis Centre