i
Kyndryl
Proud winner of ABECA 2024 - AmbitionBox Employee Choice Awards
Filter interviews by
The question is about a Pyspark problem.
Use SparkSession to create a Spark application
Load data from a source like CSV or Parquet files
Perform transformations and actions on the data using PySpark functions
Optimize performance by using caching and partitioning
Top trending discussions
posted on 31 Dec 2024
Apache Spark architecture includes a cluster manager, worker nodes, and driver program.
Apache Spark architecture consists of a cluster manager, which allocates resources and schedules tasks.
Worker nodes execute tasks and store data in memory or disk.
Driver program coordinates tasks and communicates with the cluster manager.
Spark applications run as independent sets of processes on a cluster, coordinated by the SparkCon...
reduceBy is used to aggregate data based on key, while groupBy is used to group data based on key.
reduceBy is a transformation that combines the values of each key using an associative function and a neutral 'zero value'.
groupBy is a transformation that groups the data based on a key and returns a grouped data set.
reduceBy is more efficient for aggregating data as it reduces the data before shuffling, while groupBy shu...
RDD is a low-level abstraction representing a distributed collection of objects, while DataFrame is a higher-level abstraction representing a distributed collection of data organized into named columns.
RDD is more suitable for unstructured data and low-level transformations, while DataFrame is more suitable for structured data and high-level abstractions.
DataFrames provide optimizations like query optimization and code...
The different modes of execution in Apache Spark include local mode, standalone mode, YARN mode, and Mesos mode.
Local mode: Spark runs on a single machine with one executor.
Standalone mode: Spark runs on a cluster managed by a standalone cluster manager.
YARN mode: Spark runs on a Hadoop cluster using YARN as the resource manager.
Mesos mode: Spark runs on a Mesos cluster with Mesos as the resource manager.
I applied via LinkedIn and was interviewed in Nov 2024. There was 1 interview round.
I applied via Campus Placement and was interviewed in Oct 2024. There was 1 interview round.
Use regular expression to remove special characters from a string
Use the regex pattern [^a-zA-Z0-9\s] to match any character that is not a letter, digit, or whitespace
Use the replace() function in your programming language to replace the matched special characters with an empty string
Example: input string 'Hello! How are you?' will become 'Hello How are you' after removing special characters
I applied via Job Portal and was interviewed in Aug 2024. There were 3 interview rounds.
Its mandatory test even for experience people
Asked me two string array question one was to reverse a string without any pre build function and second one was a medium question to print the number and the count of it into the next level of the tree
I manage my work by prioritizing tasks, setting goals, staying organized, and communicating effectively with team members.
Prioritize tasks based on deadlines and importance
Set clear goals and milestones to track progress
Stay organized with tools like project management software
Communicate effectively with team members to ensure alignment and collaboration
posted on 19 Nov 2024
I applied via Naukri.com and was interviewed in May 2024. There was 1 interview round.
1. Questions on spark basica
2. Sql coding questions
3. Java or scala basics
based on 1 interview
Interview experience
based on 4 reviews
Rating in categories
System Administrator
1.3k
salaries
| ₹2.9 L/yr - ₹12.1 L/yr |
Associate Technical Engineer
633
salaries
| ₹3.5 L/yr - ₹6.5 L/yr |
Project Manager
622
salaries
| ₹6.7 L/yr - ₹29 L/yr |
Infrastructure Specialist
556
salaries
| ₹4.1 L/yr - ₹24 L/yr |
SME
435
salaries
| ₹9.5 L/yr - ₹27 L/yr |
IBM
Accenture
TCS
Wipro