Filter interviews by
I applied via Approached by Company and was interviewed before Oct 2023. There were 3 interview rounds.
Developing a data warehouse to consolidate and analyze sales data from multiple sources
Designing ETL processes to extract, transform, and load data into the warehouse
Creating data models and reports for business users to analyze sales performance
Implementing data quality checks to ensure accuracy and consistency of the data
Working with stakeholders to gather requirements and prioritize development tasks
Challenges during deployment include data migration, system compatibility, and user training.
Data migration can be complex and time-consuming, especially when dealing with large volumes of data.
Ensuring system compatibility with existing infrastructure and software can be a challenge, requiring thorough testing and troubleshooting.
User training is essential to ensure successful adoption of the new BI system, but can be...
I applied via Company Website and was interviewed in Nov 2024. There were 3 interview rounds.
Standard quantitative questions accompanied by logical reasoning and related topics.
OOP concepts include encapsulation, inheritance, polymorphism, and abstraction.
Encapsulation: Bundling data and methods that operate on the data into a single unit (class).
Inheritance: Allows a class to inherit properties and behavior from another class.
Polymorphism: Objects of different classes can be treated as objects of a common superclass.
Abstraction: Hides complex implementation details and only shows the necessa
React is a JavaScript library for building user interfaces.
React allows developers to create reusable UI components.
It uses a virtual DOM for efficient rendering.
React uses a declarative approach to define how the UI should look based on the application state.
It supports server-side rendering for improved performance.
React can be used with other libraries like Redux for state management.
I am a software engineer with a passion for coding and problem-solving. I come from a close-knit family with supportive parents and siblings.
Software engineer with coding and problem-solving skills
Close-knit family with supportive parents and siblings
My strengths include problem-solving and teamwork, while my weaknesses are time management and public speaking. I am addressing my weaknesses by taking time management courses and practicing public speaking.
Strengths: problem-solving, teamwork
Weaknesses: time management, public speaking
Addressing weaknesses: taking time management courses, practicing public speaking
I applied via campus placement at SRM Institute of Science and Technology, NCR Campus, Ghaziabad and was interviewed in Nov 2024. There was 1 interview round.
I am a passionate software engineer with 5 years of experience in developing web applications using various technologies.
5 years of experience in software development
Proficient in developing web applications
Skilled in using various technologies
Passionate about software engineering
Developed a web-based student management system for tracking attendance and grades.
Used HTML, CSS, and JavaScript for front-end development.
Implemented backend functionality using PHP and MySQL database.
Incorporated user authentication and authorization features for secure access.
Tested the system thoroughly to ensure functionality and user-friendliness.
Graduated with honors, completed multiple internships, and won coding competitions
Graduated with honors in Computer Science
Completed internships at XYZ Company and ABC Corporation
Won first place in coding competition at university hackathon
posted on 14 Dec 2024
I was interviewed in Nov 2024.
Use 'hdfs diskbalancer' command to check disk utilisation and health in Hadoop
Run 'hdfs diskbalancer -report' to get a report on disk utilisation
Use 'hdfs diskbalancer -plan <path>' to generate a plan for balancing disk usage
Check the Hadoop logs for any disk health issues
Spark Architecture consists of Driver, Cluster Manager, and Executors. Driver manages the execution of Spark jobs.
Driver: Manages the execution of Spark jobs, converts user code into tasks, and coordinates with Cluster Manager.
Cluster Manager: Manages resources across the cluster and allocates resources to Spark applications.
Executors: Execute tasks assigned by the Driver and store data in memory or disk for further pr...
Optimization techniques in Spark improve performance and efficiency of data processing.
Partitioning data to distribute workload evenly
Caching frequently accessed data in memory
Using broadcast variables for small lookup tables
Avoiding shuffling operations whenever possible
Tuning memory settings and garbage collection parameters
I am unable to provide this information as it is confidential.
Confidential information about salaries in previous organizations should not be disclosed.
It is important to respect the privacy and confidentiality of past employers.
Discussing specific salary details may not be appropriate in a professional setting.
To create a pivot table in SQL from a non-pivot table, you can use the CASE statement with aggregate functions.
Use the CASE statement to categorize data into columns
Apply aggregate functions like SUM, COUNT, AVG, etc. to calculate values for each category
Group the data by the columns you want to pivot on
Creating triggers in a database involves defining the trigger, specifying the event that will activate it, and writing the code to be executed.
Define the trigger using the CREATE TRIGGER statement
Specify the event that will activate the trigger (e.g. INSERT, UPDATE, DELETE)
Write the code or actions to be executed when the trigger is activated
Test the trigger to ensure it functions as intended
I applied via Campus Placement and was interviewed in Oct 2024. There were 5 interview rounds.
Basic aptitudes question in online test
Two coding Q online test
I want to join Global Logic because of their reputation for cutting-edge technology solutions and collaborative work environment.
Global Logic is known for their innovative technology solutions
I am impressed by the collaborative work environment at Global Logic
I believe Global Logic offers great opportunities for professional growth and development
I applied via Naukri.com and was interviewed in Aug 2024. There were 2 interview rounds.
Different types of joins in SQL with examples
Inner Join: Returns rows when there is a match in both tables
Left Join: Returns all rows from the left table and the matched rows from the right table
Right Join: Returns all rows from the right table and the matched rows from the left table
Full Outer Join: Returns all rows when there is a match in either table
Large Spark datasets can be handled by partitioning, caching, optimizing transformations, and tuning resources.
Partitioning data to distribute workload evenly across nodes
Caching frequently accessed data to avoid recomputation
Optimizing transformations to reduce unnecessary processing
Tuning resources like memory allocation and parallelism for optimal performance
Spark configuration settings can be tuned to optimize query performance by adjusting parameters like memory allocation, parallelism, and caching.
Increase executor memory and cores to allow for more parallel processing
Adjust shuffle partitions to optimize data shuffling during joins and aggregations
Enable dynamic allocation to scale resources based on workload demands
Utilize caching to store intermediate results and avo...
To handle data skew and partition imbalance in Spark, strategies include using salting, bucketing, repartitioning, and optimizing join operations.
Use salting to evenly distribute skewed keys across partitions
Implement bucketing to pre-partition data based on a specific column
Repartition data based on a specific key to balance partitions
Optimize join operations by broadcasting small tables or using partitioning strategi
To create a custom serializable interface, you need to define an interface with the Serializable marker interface and implement the necessary methods.
Define an interface with the Serializable marker interface
Implement the necessary methods for serialization and deserialization
Ensure all fields in the class implementing the interface are serializable
Criteria in Hibernate is used to create and execute dynamic queries.
Criteria is an interface in Hibernate that allows the creation of dynamic queries without using HQL or SQL.
Criteria queries are type-safe and can be easily modified at runtime.
Criteria queries can be used to fetch entities based on certain conditions or restrictions.
Example: Criteria criteria = session.createCriteria(Employee.class);
I applied via Referral and was interviewed in Oct 2024. There were 2 interview rounds.
Dominations of coins it was a dp question
Implement a stack that supports finding the minimum element in constant time.
Use two stacks - one to store the actual elements and another to store the minimum values at each level.
When pushing an element, compare it with the current minimum and push the smaller value onto the minimum stack.
When popping an element, also pop from the minimum stack if the popped element is the current minimum.
LCA (Lowest Common Ancestor) of given nodes in a tree is the node that is the lowest common ancestor of the given nodes.
Traverse the tree to find the paths from the root to each of the given nodes.
Compare the paths to find the last common node, which is the LCA.
If one of the given nodes is the ancestor of the other, return the ancestor node as the LCA.
I applied via campus placement at Noida Institute of Engineering & Technology, Greater Noida and was interviewed in Oct 2024. There were 3 interview rounds.
The topic given to us was "Can AI replace human Intelligence?"
I applied via campus placement at Dronacharya College of Engineering, Gurgaon and was interviewed in Nov 2024. There were 2 interview rounds.
Three coding question
Two technical questions based on strings and linked lists.
Senior Software Engineer
3
salaries
| ₹14.5 L/yr - ₹21 L/yr |
Greenstone
Invenio Business Solutions
Accenture
Wipro