Filter interviews by
Clear (1)
I applied via LinkedIn and was interviewed in Jan 2024. There was 1 interview round.
Pyspark is a Python API for Apache Spark, a powerful open-source distributed computing system.
Pyspark is used for processing large datasets in parallel across a cluster of computers.
It provides high-level APIs in Python for Spark programming.
Pyspark allows seamless integration with other Python libraries like Pandas and NumPy.
Example: Using Pyspark to perform data analysis and machine learning tasks on big data sets.
Pyspark SQL is a module in Apache Spark that provides a SQL interface for working with structured data.
Pyspark SQL allows users to run SQL queries on Spark dataframes.
It provides a more concise and user-friendly way to interact with data compared to traditional Spark RDDs.
Users can leverage the power of SQL for data manipulation and analysis within the Spark ecosystem.
To merge 2 dataframes of different schema, use join operations or data transformation techniques.
Use join operations like inner join, outer join, left join, or right join based on the requirement.
Perform data transformation to align the schemas before merging.
Use tools like Apache Spark, Pandas, or SQL to merge dataframes with different schemas.
Pyspark streaming is a scalable and fault-tolerant stream processing engine built on top of Apache Spark.
Pyspark streaming allows for real-time processing of streaming data.
It provides high-level APIs in Python for creating streaming applications.
Pyspark streaming supports various data sources like Kafka, Flume, Kinesis, etc.
It enables windowed computations and stateful processing for handling streaming data.
Example: C...
I applied via Company Website and was interviewed in Jan 2024. There was 1 interview round.
Spark architecture includes driver, cluster manager, and worker nodes for distributed processing.
Spark architecture consists of a driver program that manages the execution of tasks on worker nodes.
Cluster manager is responsible for allocating resources and scheduling tasks across worker nodes.
Worker nodes execute the tasks and store data in memory or disk for processing.
Example: In a Spark application, the driver progr...
I applied via Recruitment Consulltant and was interviewed before Jul 2023. There were 2 interview rounds.
Handling ADF pipelines involves designing, building, and monitoring data pipelines in Azure Data Factory.
Designing data pipelines using ADF UI or code
Building pipelines with activities like copy data, data flow, and custom activities
Monitoring pipeline runs and debugging issues
Optimizing pipeline performance and scheduling triggers
What people are saying about Luxoft
Technical questions from hive , spark Scala and azure
Luxoft interview questions for designations
I was interviewed before Sep 2020.
Round duration - 140 minutes
Round difficulty - Medium
Test timing was at 2:00 pm , it was conducted in a college and the environment was good for the test. Camera was a primary part of test, so no suspicious activity.
Given two numbers in the form of two arrays where each element of the array represents a digit, calculate the sum of these two numbers and return this sum as an ar...
Given two numbers represented as arrays, calculate their sum and return the result as an array.
Iterate through the arrays from right to left, adding digits and carrying over if necessary
Handle cases where one array is longer than the other by considering the remaining digits
Ensure the final sum array does not have any leading zeros
Round duration - 20 minutes
Round difficulty - Easy
The round was conducted at around 12 p.m. I was called at the college location and then it was conducted. The interviewer was quite polite and frank.
Round duration - 8 minutes
Round difficulty - Easy
This round was conducted right after finishing and clearing the technical round at the same place and on the same day.
Tip 1 : Practice atleast 2-3 Coding problems daily so your logic building becomes stronger.
Tip 2 : Exercise problems based on OOPS concepts and others too.
Tip 3 : If you can have your own project built, then it's the major point and will act as a plus point.
Tip 1 : Your resume should be in standard form, short and simple will be more effective.
Tip 2 : Whatever you have learned, you need to mention it in your resume as that will be your primary source of selection and having project on your resume is important.
I applied via Campus Placement and was interviewed before Jun 2020. There were 3 interview rounds.
I applied via Naukri.com and was interviewed before Mar 2021. There were 3 interview rounds.
Yes, I have technical certifications and I am proficient in multiple programming languages and development tools.
I have certifications in Java and Python
I am proficient in Java, Python, C++, and JavaScript
I have used development tools such as Eclipse, Visual Studio Code, and PyCharm
I applied via Campus Placement and was interviewed before May 2020. There were 3 interview rounds.
I applied via Company Website and was interviewed before Oct 2019. There were 4 interview rounds.
based on 3 interviews
1 Interview rounds
based on 1 review
Rating in categories
Senior Software Engineer
466
salaries
| ₹0 L/yr - ₹0 L/yr |
Senior Consultant
354
salaries
| ₹0 L/yr - ₹0 L/yr |
Consultant
284
salaries
| ₹0 L/yr - ₹0 L/yr |
Software Engineer
206
salaries
| ₹0 L/yr - ₹0 L/yr |
Senior Software Developer
135
salaries
| ₹0 L/yr - ₹0 L/yr |
Accenture
EPAM Systems
GlobalLogic
LTIMindtree