Filter interviews by
I appeared for an interview in Nov 2024.
Use 'hdfs diskbalancer' command to check disk utilisation and health in Hadoop
Run 'hdfs diskbalancer -report' to get a report on disk utilisation
Use 'hdfs diskbalancer -plan <path>' to generate a plan for balancing disk usage
Check the Hadoop logs for any disk health issues
Spark Architecture consists of Driver, Cluster Manager, and Executors. Driver manages the execution of Spark jobs.
Driver: Manages the execution of Spark jobs, converts user code into tasks, and coordinates with Cluster Manager.
Cluster Manager: Manages resources across the cluster and allocates resources to Spark applications.
Executors: Execute tasks assigned by the Driver and store data in memory or disk for further pr...
Optimization techniques in Spark improve performance and efficiency of data processing.
Partitioning data to distribute workload evenly
Caching frequently accessed data in memory
Using broadcast variables for small lookup tables
Avoiding shuffling operations whenever possible
Tuning memory settings and garbage collection parameters
I am unable to provide this information as it is confidential.
Confidential information about salaries in previous organizations should not be disclosed.
It is important to respect the privacy and confidentiality of past employers.
Discussing specific salary details may not be appropriate in a professional setting.
To create a pivot table in SQL from a non-pivot table, you can use the CASE statement with aggregate functions.
Use the CASE statement to categorize data into columns
Apply aggregate functions like SUM, COUNT, AVG, etc. to calculate values for each category
Group the data by the columns you want to pivot on
Creating triggers in a database involves defining the trigger, specifying the event that will activate it, and writing the code to be executed.
Define the trigger using the CREATE TRIGGER statement
Specify the event that will activate the trigger (e.g. INSERT, UPDATE, DELETE)
Write the code or actions to be executed when the trigger is activated
Test the trigger to ensure it functions as intended
I applied via Referral and was interviewed in Mar 2022. There was 1 interview round.
Spark optimization techniques improve performance and efficiency of Spark applications.
Partitioning data to reduce shuffling
Caching frequently used data
Using broadcast variables for small data
Using efficient data formats like Parquet
Tuning memory and CPU usage
Using appropriate cluster size
Avoiding unnecessary data shuffling
Using appropriate serialization formats
Using appropriate join strategies
We use Hadoop Distributed File System (HDFS) for our project.
HDFS is a distributed file system designed to run on commodity hardware.
It provides high-throughput access to application data and is fault-tolerant.
HDFS is used by many big data processing frameworks like Hadoop, Spark, etc.
It stores data in a distributed manner across multiple nodes in a cluster.
HDFS is optimized for large files and sequential reads and wri
Developed a web-based project management tool for a startup
Used React for the frontend and Node.js for the backend
Implemented user authentication and authorization using JWT
Integrated with third-party APIs such as Trello and Slack
Implemented real-time updates using WebSockets
Deployed on AWS using EC2 and RDS
Fibonacci series is a sequence of numbers where each number is the sum of the two preceding ones.
Start with 0 and 1 as the first two numbers
Add the previous two numbers to get the next number
Repeat until desired number of terms
I am a software engineer with experience in developing web applications and mobile apps.
Proficient in programming languages such as Java, Python, and JavaScript
Skilled in using frameworks like React, Angular, and Spring Boot
Experienced in working with databases such as MySQL and MongoDB
Familiar with Agile development methodologies and DevOps practices
There are numerous ways to use a box full of pencils depending on the context and purpose.
Pencils can be used for writing, drawing, sketching, coloring, shading, and more.
They can be used for artistic purposes, schoolwork, office work, or personal use.
Pencils can also be used as a tool for measurement, as a makeshift stylus, or even as a musical instrument.
The number of ways to use a box full of pencils is virtually li...
My dream is to build innovative software solutions that positively impact people's lives.
Developing cutting-edge technology
Creating user-friendly interfaces
Solving complex problems
Collaborating with talented individuals
Making a difference in society
Continuous learning and growth
I applied via Naukri.com and was interviewed before Nov 2020. There were 3 interview rounds.
I applied via AmbitionBox and was interviewed before Jul 2021. There were 3 interview rounds.
Generally the aptitude maintain quick calculations and time reduce.
Coding built over technical skills
I applied via Indeed and was interviewed before Jun 2021. There were 4 interview rounds.
It was easy aptitude test , can be cracked easily
based on 1 interview
Interview experience
based on 3 reviews
Rating in categories
Senior Consultant
3.8k
salaries
| ₹8 L/yr - ₹30 L/yr |
Lead Consultant
3.5k
salaries
| ₹10.5 L/yr - ₹34.6 L/yr |
Software Engineer
3.3k
salaries
| ₹3.5 L/yr - ₹13.3 L/yr |
Consultant
3.3k
salaries
| ₹6 L/yr - ₹20 L/yr |
Associate Consultant
2.7k
salaries
| ₹4.7 L/yr - ₹15.1 L/yr |
Cognizant
TCS
Infosys
Accenture