Filter interviews by
I applied via Naukri.com and was interviewed in Feb 2023. There were 3 interview rounds.
Big data refers to large volumes of structured and unstructured data that is too complex for traditional data processing applications.
Big data is characterized by the 3 Vs - volume, velocity, and variety.
Volume refers to the sheer amount of data being generated, often in terabytes or petabytes.
Velocity refers to the speed at which data is being generated and processed, often in real-time.
Variety refers to the different...
I will identify and delete unnecessary files, compress data, and optimize storage to free up space.
Identify and delete unnecessary files and logs
Compress data to save space
Optimize storage by removing temporary files or old backups
The command for checking live processes on a local level is 'ps'
Use the 'ps' command to display information about currently running processes
Add options like 'aux' to show more detailed information
Use 'grep' to filter the output for specific processes
NACL is a stateless firewall that controls traffic at the subnet level, while security groups are stateful firewalls that control traffic at the instance level.
NACL operates at the subnet level, while security groups operate at the instance level.
NACL is stateless, meaning it does not keep track of the state of connections, while security groups are stateful and keep track of connections.
NACL rules are evaluated in ord...
Top trending discussions
I applied via Naukri.com and was interviewed in Sep 2022. There were 4 interview rounds.
I am a Hadoop Administrator with experience in managing and optimizing Hadoop clusters.
Experienced Hadoop Administrator
Proficient in managing and optimizing Hadoop clusters
Skilled in troubleshooting and resolving Hadoop-related issues
Knowledgeable in Hadoop ecosystem components like HDFS, MapReduce, YARN, Hive, Pig, etc.
Familiar with Hadoop security and authentication mechanisms
Experience in capacity planning and perfo
I have 3 years of experience in Big Data Administration.
I have worked with Hadoop, Spark, and Hive.
I have experience in setting up and maintaining Hadoop clusters.
I have worked with various Big Data tools and technologies.
I have experience in troubleshooting and resolving issues related to Big Data systems.
My notice period is 30 days.
I have a notice period of 30 days.
I am required to give a 30-day notice before leaving my current position.
I will be available to start a new job after the completion of my notice period.
As a Hadoop Administrator, my day-to-day activities involve managing and maintaining Hadoop clusters, monitoring system performance, troubleshooting issues, and implementing security measures.
Managing and maintaining Hadoop clusters
Monitoring system performance and resource utilization
Troubleshooting and resolving issues related to Hadoop infrastructure
Implementing security measures to protect data and ensure complianc...
The challenging aspects of a Hadoop Administrator job role
Managing and troubleshooting large-scale Hadoop clusters
Ensuring high availability and performance of the Hadoop ecosystem
Implementing security measures to protect sensitive data
Keeping up with the rapidly evolving Hadoop ecosystem and its related technologies
Optimizing resource utilization and capacity planning
Collaborating with cross-functional teams to addres...
Our project architecture follows a distributed model with Hadoop clusters for data storage and processing.
We use Hadoop clusters for distributed storage and processing of large datasets.
Our architecture includes multiple nodes with Hadoop Distributed File System (HDFS) for data storage.
We use YARN for resource management and job scheduling.
We also use Apache Spark for data processing and analysis.
Our architecture is sc...
Spark architecture is based on master-slave architecture with a cluster manager and worker nodes.
Spark has a master node that manages the cluster and worker nodes that execute tasks.
The cluster manager allocates resources to worker nodes and monitors their health.
Spark uses a distributed file system like HDFS to store data and share it across the cluster.
Spark applications are written in high-level languages like Scala...
I handled a critical incident where a Hadoop cluster crashed due to a hardware failure.
Identified the root cause of the hardware failure
Coordinated with the hardware vendor to replace the faulty components
Restored the Hadoop cluster from the latest backup
Performed thorough testing to ensure data integrity and cluster stability
We use JIRA as our ticketing tool.
JIRA is a popular ticketing tool used for issue tracking and project management.
It allows us to create, assign, and track issues and tasks.
We can also set priorities, due dates, and add comments and attachments.
JIRA integrates well with other tools like Confluence and Bitbucket.
We use JIRA to manage our Hadoop cluster and track any issues or bugs.
It helps us to prioritize and resolve i
Prioritize tickets based on severity and impact on business operations.
Assess the severity of the issue reported in the ticket
Determine the impact of the issue on business operations
Prioritize tickets with high severity and high impact on business operations
Ensure timely resolution of critical issues to minimize business impact
Yes, I have led a team.
Led a team of 5 Hadoop administrators in managing and maintaining a large Hadoop cluster
Assigned tasks, set goals, and provided guidance to team members
Conducted regular team meetings to discuss progress, challenges, and improvements
Implemented effective communication channels to ensure smooth collaboration within the team
Successfully completed several projects on time and within budget with the
I applied via Campus Placement and was interviewed in Oct 2024. There were 5 interview rounds.
Basic aptitudes question in online test
Two coding Q online test
I want to join Global Logic because of their reputation for cutting-edge technology solutions and collaborative work environment.
Global Logic is known for their innovative technology solutions
I am impressed by the collaborative work environment at Global Logic
I believe Global Logic offers great opportunities for professional growth and development
To create a custom serializable interface, you need to define an interface with the Serializable marker interface and implement the necessary methods.
Define an interface with the Serializable marker interface
Implement the necessary methods for serialization and deserialization
Ensure all fields in the class implementing the interface are serializable
Criteria in Hibernate is used to create and execute dynamic queries.
Criteria is an interface in Hibernate that allows the creation of dynamic queries without using HQL or SQL.
Criteria queries are type-safe and can be easily modified at runtime.
Criteria queries can be used to fetch entities based on certain conditions or restrictions.
Example: Criteria criteria = session.createCriteria(Employee.class);
Basic MCQs on General aptitude, logical reasoning and programming questions
A constructor in C++ is a special member function that is automatically called when an object is created.
Constructors have the same name as the class and do not have a return type.
They can be used to initialize the object's data members.
Default constructors are automatically generated if no constructor is defined.
Parameterized constructors can be defined to accept arguments during object creation.
Copy assignment operator is a special member function in C++ that allows one object to be assigned the values of another object of the same class.
Copy assignment operator is denoted by the = operator.
It is used to copy the values of one object into another object of the same class.
It is automatically generated by the compiler if not explicitly defined.
Example: MyClass obj1; MyClass obj2; obj2 = obj1; // Copy assignment
Virtualization in C++ is the process of creating a virtual version of a class or function to allow for polymorphism and dynamic binding.
Virtualization in C++ is achieved through the use of virtual functions and inheritance.
Virtual functions are declared in a base class and overridden in derived classes to provide different implementations.
Virtualization allows for polymorphism, where a pointer to a base class can point...
Implement a custom String class in a programming language.
Define a class with necessary properties and methods to manipulate strings.
Include methods for concatenation, substring, length, etc.
Handle memory allocation and deallocation properly.
Example: class MyString { // implementation }
Smart pointers are objects that act like pointers but provide automatic memory management.
Smart pointers help prevent memory leaks by automatically deallocating memory when no longer needed.
Examples include unique_ptr, shared_ptr, and weak_ptr in C++.
They provide better memory safety compared to raw pointers.
I applied via Recruitment Consulltant and was interviewed in Sep 2024. There were 3 interview rounds.
30 questions combination of Logical, numeric, verbal and 2 coding questions.
I am a passionate software engineer with experience in developing web applications and a strong background in computer science.
Experienced in developing web applications using technologies such as HTML, CSS, JavaScript, and React
Strong background in computer science with knowledge of data structures and algorithms
Proficient in programming languages such as Java, Python, and C++
Familiar with Agile development methodolog
Strength: Problem-solving skills. Weakness: Overthinking.
Strength: I have strong problem-solving skills which help me in finding efficient solutions to complex issues.
Weakness: Sometimes I tend to overthink situations, which can lead to unnecessary stress and delays in decision-making.
I applied via campus placement at Dronacharya College of Engineering, Gurgaon and was interviewed in Nov 2024. There were 2 interview rounds.
Three coding question
Two technical questions based on strings and linked lists.
I applied via Company Website and was interviewed in Oct 2024. There was 1 interview round.
posted on 7 Jan 2025
I applied via Job Portal and was interviewed in Jul 2024. There were 2 interview rounds.
What is apex and trigger
asked on lwc
Integration using rest api
Best practices
I applied via Naukri.com and was interviewed in Aug 2024. There was 1 interview round.
Senior Engineer
5.3k
salaries
| ₹5 L/yr - ₹17 L/yr |
Engineer
4.7k
salaries
| ₹1.5 L/yr - ₹8.5 L/yr |
Technical Lead
2k
salaries
| ₹8.5 L/yr - ₹30 L/yr |
Project Lead
1.6k
salaries
| ₹6 L/yr - ₹23 L/yr |
Senior Software Engineer
1.4k
salaries
| ₹4.8 L/yr - ₹18.2 L/yr |
TCS
Infosys
Wipro
Tech Mahindra