i
Filter interviews by
L1 and L2 regularizations are techniques to prevent overfitting in machine learning models by adding penalties to the loss function.
L1 regularization (Lasso) adds the absolute value of coefficients to the loss function, promoting sparsity.
L2 regularization (Ridge) adds the squared value of coefficients, which discourages large weights but retains all features.
L1 can lead to feature selection by driving some coeffi...
Overfitting occurs when a model learns the training data too well, leading to poor generalization. Underfitting happens when a model is too simple to capture the underlying patterns.
Overfitting: Model performs well on training data but poorly on unseen data. Can be caused by a model being too complex or training for too long.
Underfitting: Model is too simple to capture the underlying patterns in the data. Results ...
To overcome overfitting, use techniques like cross-validation, regularization, early stopping, and increasing training data.
Use cross-validation to evaluate model performance on different subsets of data.
Apply regularization techniques like L1 or L2 regularization to penalize large coefficients.
Implement early stopping to stop training when validation error starts to increase.
Increase training data to provide more...
Transformers are a type of neural network architecture designed for processing sequential data, particularly in NLP tasks.
Transformers use self-attention mechanisms to weigh the importance of different words in a sentence.
They consist of an encoder-decoder structure, where the encoder processes input data and the decoder generates output.
Transformers can handle long-range dependencies better than RNNs due to their...
Transformer architecture is a neural network architecture used for natural language processing tasks. Positional encoders are used to encode the position of words in a sentence.
Transformer architecture is based on the self-attention mechanism.
It consists of an encoder and a decoder.
Positional encoders are added to the input embeddings to encode the position of words in a sentence.
They are computed using sine and c...
ML algorithms are techniques used to train models to make predictions or decisions based on data.
Supervised learning algorithms: Linear regression, logistic regression, decision trees, random forests, support vector machines, k-nearest neighbors
Unsupervised learning algorithms: K-means clustering, hierarchical clustering, principal component analysis
Reinforcement learning algorithms: Q-learning, SARSA
Deep learning...
Different types of learning include supervised learning, unsupervised learning, semi-supervised learning, reinforcement learning, and transfer learning.
Supervised learning: Training a model using labeled data to make predictions or classifications.
Unsupervised learning: Training a model on unlabeled data to discover patterns or relationships.
Semi-supervised learning: Combining labeled and unlabeled data for traini...
Naive Bayes is a probabilistic algorithm that uses Bayes' theorem to classify data based on prior knowledge.
Naive Bayes assumes that all features are independent of each other.
It is commonly used for text classification and spam filtering.
There are three types of Naive Bayes classifiers: Gaussian, Multinomial, and Bernoulli.
It is a fast and simple algorithm that works well with high-dimensional datasets.
Naive Baye...
Regression is a statistical method used to analyze the relationship between a dependent variable and one or more independent variables.
Regression is used to predict continuous numerical values.
It helps in identifying the strength and direction of the relationship between variables.
Linear regression is a common type of regression used to model the relationship between two variables.
Examples of regression include pr...
PCA is a dimensionality reduction technique used to reduce the number of features in a dataset while preserving the most important information.
PCA stands for Principal Component Analysis
It works by finding the directions (principal components) in which the data varies the most
These principal components are orthogonal to each other and capture the maximum variance in the data
Feature selection can be done by selecti...
I applied via LinkedIn and was interviewed in Apr 2024. There were 2 interview rounds.
To overcome overfitting, use techniques like cross-validation, regularization, early stopping, and increasing training data.
Use cross-validation to evaluate model performance on different subsets of data.
Apply regularization techniques like L1 or L2 regularization to penalize large coefficients.
Implement early stopping to stop training when validation error starts to increase.
Increase training data to provide more dive...
PCA is a dimensionality reduction technique used to reduce the number of features in a dataset while preserving the most important information.
PCA stands for Principal Component Analysis
It works by finding the directions (principal components) in which the data varies the most
These principal components are orthogonal to each other and capture the maximum variance in the data
Feature selection can be done by selecting th...
I applied via Campus Placement and was interviewed in Jun 2024. There were 2 interview rounds.
Contain question related to aptitude, pyrhon,ml mcqs.
Overfitting occurs when a model learns the training data too well, leading to poor generalization. Underfitting happens when a model is too simple to capture the underlying patterns.
Overfitting: Model performs well on training data but poorly on unseen data. Can be caused by a model being too complex or training for too long.
Underfitting: Model is too simple to capture the underlying patterns in the data. Results in po...
I applied via Campus Placement
78 mcq questions with 2 coding questions in 1 hr 26 min.
Aptitude round - Quant reasoning verbal
Three coding questions were given
I applied via Campus Placement and was interviewed in Dec 2023. There were 2 interview rounds.
Big data, aptitude, html, css, javascript, c language questions
Hackerrank coding test
I appeared for an interview before Jun 2024, where I was asked the following questions.
L1 and L2 regularizations are techniques to prevent overfitting in machine learning models by adding penalties to the loss function.
L1 regularization (Lasso) adds the absolute value of coefficients to the loss function, promoting sparsity.
L2 regularization (Ridge) adds the squared value of coefficients, which discourages large weights but retains all features.
L1 can lead to feature selection by driving some coefficient...
Gradient descent is an optimization algorithm used to minimize a function by iteratively moving towards the steepest descent.
Gradient descent updates parameters in the opposite direction of the gradient of the loss function.
The learning rate determines the size of the steps taken towards the minimum.
There are different variants: Batch Gradient Descent, Stochastic Gradient Descent (SGD), and Mini-batch Gradient Descent.
...
I applied via Campus Placement and was interviewed before Oct 2022. There were 4 interview rounds.
OOP, aptitude, DSA, coding question-2 mcq and 1 code based
OOPs stands for Object-Oriented Programming and its 4 pillars are Inheritance, Encapsulation, Abstraction, and Polymorphism.
Inheritance allows a class to inherit properties and behavior from another class.
Encapsulation restricts access to certain components of an object, protecting its integrity.
Abstraction hides complex implementation details and only shows the necessary features.
Polymorphism allows objects to be trea...
I applied via Campus Placement and was interviewed before Sep 2022. There were 4 interview rounds.
Questions on Quantitative, English and Logic.
Next section comprised of questions on SQL, OS, Java, HTML/CSS.
Last section had questions related to python and Machine learning and also couple of Coding questions whose level was Easy to moderate.
Explanation of evaluation metrics, precision, recall, entropy, Gini impurity, bagging, boosting, validation vs test data, LSTM, GRU, K-means clustering, and importing CSV datasets.
Evaluation metrics: used to measure the performance of machine learning models (e.g., accuracy, precision, recall, F1 score)
Precision: ratio of true positive predictions to the total predicted positives (TP / (TP + FP))
Recall: ratio of true p...
I applied via Superset and was interviewed before May 2022. There were 2 interview rounds.
I applied via Campus Placement and was interviewed before Apr 2023. There was 1 interview round.
Top trending discussions
Some of the top questions asked at the Quantiphi Analytics Solutions Private Limited Machine Learning Engineer interview -
The duration of Quantiphi Analytics Solutions Private Limited Machine Learning Engineer interview process can vary, but typically it takes about 2-4 weeks to complete.
based on 11 interview experiences
Difficulty level
Duration
based on 63 reviews
Rating in categories
Data Engineer
520
salaries
| ₹6.4 L/yr - ₹15.2 L/yr |
Senior Data Engineer
450
salaries
| ₹12.3 L/yr - ₹23.5 L/yr |
Senior Business Analyst
365
salaries
| ₹12.4 L/yr - ₹22 L/yr |
Machine Learning Engineer
327
salaries
| ₹5.5 L/yr - ₹13.4 L/yr |
Business Analyst
325
salaries
| ₹6.5 L/yr - ₹14 L/yr |
ITC Infotech
CMS IT Services
KocharTech
3i Infotech