Faster and better experience!
i
Encora
Filter interviews by
I applied via Approached by Company and was interviewed in Mar 2024. There were 4 interview rounds.
I applied via Naukri.com and was interviewed before Nov 2023. There were 2 interview rounds.
Types of structured query languages include SQL, PL/SQL, T-SQL, and others.
SQL (Structured Query Language) - widely used for managing relational databases
PL/SQL (Procedural Language/SQL) - Oracle's proprietary extension for SQL
T-SQL (Transact-SQL) - Microsoft's extension for SQL used in SQL Server
Others - include languages like MySQL, PostgreSQL, SQLite, etc.
Types of joins in SQL include inner join, left join, right join, and full outer join.
Inner join: Returns rows when there is a match in both tables
Left join: Returns all rows from the left table and the matched rows from the right table
Right join: Returns all rows from the right table and the matched rows from the left table
Full outer join: Returns all rows when there is a match in either table
I applied via Naukri.com and was interviewed in Dec 2024. There was 1 interview round.
I got several invitation calls from 3 different persons for the same interview at Xebia Bangalore Brigade office.I attended an interview at Xebia on January 11, 2025, and the experience was disappointing. Despite reading several negative reviews beforehand, I chose to give the company a fair chance, but unfortunately, the concerns expressed in those reviews turned out to be valid.
From the very beginning, the process was poorly managed. I waited for over three hours before being called, while candidates who arrived after me were invited for their interviews earlier. This inconsistency immediately raised questions about the fairness of their process.
When my turn finally came, the interview began with a moderately challenging SQL question: I was asked to fetch all invalid December month transaction IDs (which is coming in ooo hours) from a dataset, applying conditions such as working hours from Monday to Friday (9 AM to 4 PM), excluding weekends and specific holidays (24th and 25th December). While I attempted to solve this, the interviewer interrupted repeatedly with casual, unrelated remarks. These interruptions disrupted my concentration and added unnecessary pressure, making it difficult to focus on solving the query effectively.
Following this, the interviewer moved to a Python question, which involved determining whether a given number was a perfect square. Although the problem itself was simple, it included irrelevant details, such as pre-imported libraries in a web-based IDE. This added an unnecessary layer of complexity and confusion. Again, the interviewer’s interruptions and casual talk distracted me further. Instead of focusing on assessing my logic and problem-solving skills, he seemed more interested in making irrelevant comments.
What stood out most negatively was the interviewer’s unprofessional behavior. At one point, he made an inappropriate remark about my name, comparing it to his own, which he claimed was not as "weighted."
I asked his name politely and he replied " Vaibhav Gupta"
While I attempted to steer the conversation back to technical discussions, his attitude remained dismissive and unfocused. He even questioned my leadership skills but turned it into an argument instead of allowing me to explain.
I also noticed disparities in how candidates were treated. For instance, a female candidate before me was given over an hour for her interview, while mine felt rushed and dismissive. While this is my personal observation, it raised concerns about bias in their evaluation process.
The interview ended abruptly and on a negative note. When I tried to discuss architectural patterns for data pipelines, the interviewer dismissed my points outright, stating that they did not need data architects. Without providing proper closure, he left the room, leaving me feeling disrespected and undervalued.
Overall, the experience was frustrating and insulting. The interviewer’s behavior was unprofessional and dismissive, and the process lacked the basic respect and fairness expected in a professional setting. Based on my experience, I strongly believe that Xebia needs to overhaul their interview practices, ensuring a more structured, unbiased, and respectful approach toward candidates.
I am relieved I was not selected, as this experience highlighted what could likely be a toxic work environment. I would not recommend Xebia to anyone, as their lack of professionalism and courtesy reflects poorly on their organizational culture.
I applied via Naukri.com and was interviewed in Oct 2024. There were 2 interview rounds.
I was interviewed in Aug 2024.
I applied via Naukri.com and was interviewed in Oct 2024. There was 1 interview round.
Incremental load in pyspark refers to loading only new or updated data into a dataset without reloading the entire dataset.
Use the 'delta' function in pyspark to perform incremental loads by specifying the 'mergeSchema' option.
Utilize the 'partitionBy' function to optimize incremental loads by partitioning the data based on specific columns.
Implement a logic to identify new or updated records based on timestamps or uni...
I applied via LinkedIn and was interviewed in Jan 2024. There was 1 interview round.
Pyspark is a Python API for Apache Spark, a powerful open-source distributed computing system.
Pyspark is used for processing large datasets in parallel across a cluster of computers.
It provides high-level APIs in Python for Spark programming.
Pyspark allows seamless integration with other Python libraries like Pandas and NumPy.
Example: Using Pyspark to perform data analysis and machine learning tasks on big data sets.
Pyspark SQL is a module in Apache Spark that provides a SQL interface for working with structured data.
Pyspark SQL allows users to run SQL queries on Spark dataframes.
It provides a more concise and user-friendly way to interact with data compared to traditional Spark RDDs.
Users can leverage the power of SQL for data manipulation and analysis within the Spark ecosystem.
To merge 2 dataframes of different schema, use join operations or data transformation techniques.
Use join operations like inner join, outer join, left join, or right join based on the requirement.
Perform data transformation to align the schemas before merging.
Use tools like Apache Spark, Pandas, or SQL to merge dataframes with different schemas.
Pyspark streaming is a scalable and fault-tolerant stream processing engine built on top of Apache Spark.
Pyspark streaming allows for real-time processing of streaming data.
It provides high-level APIs in Python for creating streaming applications.
Pyspark streaming supports various data sources like Kafka, Flume, Kinesis, etc.
It enables windowed computations and stateful processing for handling streaming data.
Example: C...
I was interviewed in Jan 2024.
I have used various types of joins including inner join, left join, right join, and full outer join.
Used inner join to retrieve records that have matching values in both tables
Utilized left join to retrieve all records from the left table and matching records from the right table
Employed right join to retrieve all records from the right table and matching records from the left table
Utilized full outer join to retrieve ...
Query for joins in SQL to combine data from multiple tables
Use JOIN keyword to combine data from two or more tables based on a related column
Types of joins include INNER JOIN, LEFT JOIN, RIGHT JOIN, and FULL JOIN
Example: SELECT * FROM table1 INNER JOIN table2 ON table1.id = table2.id
I applied via Referral and was interviewed in Feb 2024. There was 1 interview round.
Just focus on the basics of pyspark.
based on 6 reviews
Rating in categories
Senior Software Engineer
746
salaries
| ₹9.9 L/yr - ₹36 L/yr |
Technical Lead
737
salaries
| ₹15 L/yr - ₹46 L/yr |
Software Engineer
269
salaries
| ₹4.5 L/yr - ₹16.8 L/yr |
Software Developer
148
salaries
| ₹3 L/yr - ₹13.1 L/yr |
Senior Software Developer
126
salaries
| ₹8.7 L/yr - ₹30.7 L/yr |
TCS
Infosys
Wipro
HCLTech