Filter interviews by
I applied via Campus Placement
Big data refers to large volumes of structured and unstructured data that is too complex for traditional data processing applications.
Big data involves massive amounts of data that cannot be easily managed or analyzed using traditional methods
It includes structured data (like databases) and unstructured data (like social media posts)
Examples include analyzing customer behavior on e-commerce websites, processing sensor ...
Technologies related to big data include Hadoop, Spark, Kafka, and NoSQL databases.
Hadoop - Distributed storage and processing framework for big data
Spark - In-memory data processing engine for big data analytics
Kafka - Distributed streaming platform for handling real-time data feeds
NoSQL databases - Non-relational databases for storing and retrieving large volumes of data
Data warehousing is the process of collecting, storing, and managing data from various sources for analysis and reporting.
Data warehousing involves extracting data from multiple sources
Data is transformed and loaded into a central repository
Allows for complex queries and analysis to be performed on the data
Examples include data warehouses like Amazon Redshift, Google BigQuery
Cloud in big data refers to using cloud computing services to store, manage, and analyze large volumes of data.
Cloud computing allows for scalable and flexible storage of big data
It provides on-demand access to computing resources for processing big data
Examples include AWS, Google Cloud, and Microsoft Azure
Python is a versatile programming language used for various purposes including web development, data analysis, artificial intelligence, and automation.
Python is used for web development with frameworks like Django and Flask.
It is commonly used for data analysis and visualization with libraries like Pandas and Matplotlib.
Python is popular in artificial intelligence and machine learning projects with libraries like Tenso...
Scala is a programming language that is used for building scalable and high-performance applications.
Scala is used for developing applications that require high performance and scalability.
It is often used in Big Data processing frameworks like Apache Spark.
Scala combines object-oriented and functional programming paradigms.
It is interoperable with Java, allowing developers to leverage existing Java libraries.
Scala is ...
Top trending discussions
I applied via Naukri.com and was interviewed in Dec 2024. There was 1 interview round.
To create a pipeline in Databricks, you can use Databricks Jobs or Apache Airflow for orchestration.
Use Databricks Jobs to create a pipeline by scheduling notebooks or Spark jobs.
Utilize Apache Airflow for more complex pipeline orchestration with dependencies and monitoring.
Leverage Databricks Delta for managing data pipelines with ACID transactions and versioning.
I applied via Approached by Company and was interviewed in Aug 2024. There were 3 interview rounds.
AB testing is a method used to compare two versions of a webpage or app to determine which one performs better.
AB testing involves creating two versions (A and B) of a webpage or app with one differing element
Users are randomly assigned to either version A or B to measure performance metrics
The version that performs better in terms of the desired outcome is selected for implementation
Example: Testing two different call...
It was a classification problem
posted on 26 Jul 2024
Sql assessment round
I applied via Instahyre and was interviewed in Jun 2024. There were 3 interview rounds.
Coding round had one SQL and one Python question. And some mcqs around python and math
I applied via Referral and was interviewed in May 2024. There were 4 interview rounds.
Python and SQL questions were asked
posted on 29 May 2024
I applied via Campus Placement and was interviewed in Apr 2024. There were 2 interview rounds.
It was a written test where theoretical SQL questions were asked like primary key, foreign key, set operators and some queries
posted on 12 Feb 2024
Azure Data Lake is a cloud-based storage and analytics service, while Delta Lake is an open-source storage layer that adds reliability to data lakes.
Azure Data Lake is a service provided by Microsoft Azure for storing and analyzing large amounts of data.
Delta Lake is an open-source storage layer that adds ACID transactions and schema enforcement to data lakes.
Azure Data Lake is a cloud-based solution, while Delta Lake ...
Primary keys and foreign keys can be implemented in delta tables using constraints and references.
Primary keys can be implemented using the PRIMARY KEY constraint, which ensures that each record in the table has a unique identifier.
Foreign keys can be implemented using the FOREIGN KEY constraint, which establishes a link between two tables based on a common column.
The referenced table must have a primary key defined, a...
Exception handling in Python allows for the graceful handling of errors and prevents program crashes.
Use try-except blocks to catch and handle exceptions.
Multiple except blocks can be used to handle different types of exceptions.
The finally block is executed regardless of whether an exception occurred or not.
Exceptions can be raised using the 'raise' keyword.
Custom exceptions can be defined by creating a new class that
Star schema is a data modeling technique where a central fact table is connected to multiple dimension tables. Snowflake schema is an extension of star schema with normalized dimension tables.
Star schema is a simple and denormalized structure
It consists of a central fact table connected to multiple dimension tables
Dimension tables contain descriptive attributes
Star schema is easy to understand and query, but can lead t...
The most frequently changing data
Customer preferences
Market trends
Weather data
Stock prices
Social media trends
I applied via campus placement at Sastra University and was interviewed in Mar 2024. There were 2 interview rounds.
Well designed to test the aptitude competence of the candidate.
I applied via Campus Placement and was interviewed in May 2024. There were 2 interview rounds.
Two coding questions
based on 1 interview
Interview experience
based on 2 reviews
Rating in categories
Associate Software Engineer
39
salaries
| ₹4 L/yr - ₹15 L/yr |
Data Scientist
35
salaries
| ₹5.5 L/yr - ₹21.7 L/yr |
Senior Software Engineer
34
salaries
| ₹6 L/yr - ₹21 L/yr |
Associate Technical Specialist
23
salaries
| ₹6.3 L/yr - ₹14.8 L/yr |
Software Engineer
20
salaries
| ₹4.8 L/yr - ₹13.5 L/yr |
Fractal Analytics
Mu Sigma
Tiger Analytics
LatentView Analytics