Filter interviews by
I applied via Company Website and was interviewed in Oct 2024. There were 4 interview rounds.
AWS services used include S3, Redshift, Glue, EMR, and Lambda in a scalable and cost-effective architecture.
AWS S3 for storing large amounts of data
AWS Redshift for data warehousing and analytics
AWS Glue for ETL processes
AWS EMR for big data processing
AWS Lambda for serverless computing
Developed a real-time data processing pipeline for analyzing customer behavior
Designed and implemented data ingestion process using Apache Kafka
Utilized Apache Spark for data processing and analysis
Built data models and visualizations using tools like Tableau
Implemented machine learning algorithms for predictive analytics
Spark submit command is used to submit Spark applications to a cluster
Used to launch Spark applications on a cluster
Requires specifying the application JAR file, main class, and any arguments
Can set various configurations like memory allocation, number of executors, etc.
Example: spark-submit --class com.example.Main --master yarn --deploy-mode cluster myApp.jar arg1 arg2
Developed a real-time data processing pipeline for analyzing customer behavior
Designed and implemented data ingestion process using Apache Kafka
Utilized Apache Spark for data processing and analysis
Built data models and visualizations using tools like Tableau
Implemented machine learning algorithms for predictive analytics
To configure a cluster for 100 TB data, consider factors like storage capacity, processing power, network bandwidth, and fault tolerance.
Choose a distributed storage system like HDFS or Amazon S3 for scalability and fault tolerance.
Select high-capacity servers with sufficient RAM and CPU for processing large volumes of data.
Ensure high-speed network connections between nodes to facilitate data transfer.
Implement data r...
Our current project architecture involves a microservices-based approach with data pipelines for real-time processing.
Utilizing microservices architecture for scalability and flexibility
Implementing data pipelines for real-time processing of large volumes of data
Leveraging cloud services such as AWS or Azure for infrastructure
Using technologies like Apache Kafka for streaming data
Ensuring data quality and reliability t
Use a SQL query with a subquery to find the 2nd most ordered item in a category.
Use a subquery to rank items within each category based on the number of orders
Select the item with rank 2 within each category
Order the results by category and rank to get the 2nd most ordered item in each category
Top trending discussions
I applied via Campus Placement and was interviewed before Jan 2024. There were 2 interview rounds.
I was asked to modify the project code to evaluate my skills and knowledge.
Data science is a field that uses scientific methods, algorithms, and systems to extract knowledge and insights from structured and unstructured data.
Data science involves collecting, analyzing, and interpreting large amounts of data to solve complex problems.
It combines statistics, machine learning, and domain knowledge to uncover patterns and make predictions.
Examples include predicting customer behavior based on pas...
Machine learning is a subset of artificial intelligence that involves training algorithms to learn patterns from data.
Machine learning is a subset of artificial intelligence.
It involves training algorithms to learn patterns from data.
Examples include image recognition, natural language processing, and recommendation systems.
I applied via Referral and was interviewed in Mar 2022. There was 1 interview round.
I applied via Walk-in and was interviewed in Nov 2020. There were 3 interview rounds.
The process for manual extraction involves identifying the data source, selecting relevant data, and recording it manually.
Identify the data source
Select relevant data
Record data manually
Ensure accuracy and completeness
Verify data with source if necessary
My speed of coding documents is around 60-70 words per minute.
I use typing software to improve my speed.
I prioritize accuracy over speed.
I proofread my work to ensure quality.
I am constantly looking for ways to improve my speed and efficiency.
I applied via Recruitment Consultant and was interviewed in Feb 2021. There were 4 interview rounds.
A database analyst is responsible for designing, maintaining, and optimizing databases.
Designing and implementing databases
Ensuring data accuracy and security
Optimizing database performance
Creating reports and analyzing data
Collaborating with other teams to ensure data integration
Proficiency in SQL and database management systems
Knowledge of data warehousing and data mining
Experience with ETL processes
Ability to troubl
Data analysis involves identifying patterns, trends, and insights from data to make informed decisions.
Define the problem and identify the data needed
Clean and preprocess the data
Explore the data using statistical methods and visualizations
Develop and test hypotheses
Communicate findings and insights to stakeholders
I applied via Naukri.com and was interviewed in Nov 2020. There were 3 interview rounds.
based on 1 interview
Interview experience
Senior Software Engineer
9
salaries
| ₹20 L/yr - ₹60 L/yr |
Software Engineer
6
salaries
| ₹7 L/yr - ₹11.8 L/yr |
Senior Security Engineer
6
salaries
| ₹21 L/yr - ₹60 L/yr |
Technical Support Engineer
5
salaries
| ₹3.7 L/yr - ₹3.8 L/yr |
Senior Product Marketing Manager
4
salaries
| ₹37 L/yr - ₹80 L/yr |
Bigrock
HostGator.com
Net4
ZNetLive