Filter interviews by
I applied via LinkedIn and was interviewed in Jun 2022. There were 3 interview rounds.
Two coding questions on codility. One was easy and second medium. 10 MCQ questions on Big Data related technologies.
Code to print duplicate numbers in a list.
Iterate through the list and keep track of the count of each number using a dictionary.
Print the numbers that have a count greater than 1.
Spark can connect to Azure SQL Database using JDBC driver.
Download and install the JDBC driver for Azure SQL Database.
Set up the connection string with the appropriate credentials.
Use the JDBC API to connect Spark to Azure SQL Database.
Example: val df = spark.read.jdbc(jdbcUrl, tableName, connectionProperties)
Ensure that the firewall rules for the Azure SQL Database allow access from the Spark cluster.
Spark optimization techniques include partitioning, caching, and using appropriate transformations.
Partitioning data can improve performance by reducing shuffling.
Caching frequently used data can reduce the need for recomputation.
Transformations like filter, map, and reduceByKey can be used to optimize data processing.
Shuffling can be minimized by using operations like reduceByKey instead of groupByKey.
Broadcasting sma...
Cache and persist are used to store data in memory. Repartition and coalesce are used to change the number of partitions.
Cache stores the data in memory for faster access while persist allows the user to choose the storage level.
Repartition increases the number of partitions while coalesce decreases the number of partitions.
Cache and persist are transformations while repartition and coalesce are actions.
Cache and persi...
Hive has two types of tables - Managed and External. Managed tables are managed by Hive, while External tables are managed outside of Hive.
Managed tables are created using 'CREATE TABLE' command and data is stored in Hive's warehouse directory
External tables are created using 'CREATE EXTERNAL TABLE' command and data is stored outside of Hive's warehouse directory
Managed tables are deleted when the table is dropped, whi...
Developed a data pipeline to process and analyze customer behavior data.
Used Apache Kafka for real-time data streaming
Implemented data processing using Apache Spark
Stored data in Hadoop Distributed File System (HDFS)
Used Tableau for data visualization
Code to print reverse of a sentence word by word.
Split the sentence into words using space as delimiter
Store the words in an array
Print the words in reverse order
RDD, Dataframe, and Dataset are data structures in Apache Spark with different characteristics and functionalities.
RDD (Resilient Distributed Datasets) is a fundamental data structure in Spark that represents an immutable distributed collection of objects. It provides low-level APIs for distributed data processing and fault tolerance.
Dataframe is a distributed collection of data organized into named columns. It is simi...
I applied via LinkedIn and was interviewed in Nov 2024. There were 4 interview rounds.
My current day-to-day tasks involve designing and implementing data pipelines, optimizing data storage and retrieval, and collaborating with cross-functional teams.
Designing and implementing data pipelines to extract, transform, and load data from various sources
Optimizing data storage and retrieval processes for efficiency and scalability
Collaborating with cross-functional teams to understand data requirements and del...
The end-to-end project architecture involves designing and implementing the entire data pipeline from data ingestion to data visualization.
Data ingestion: Collecting data from various sources such as databases, APIs, and files.
Data processing: Cleaning, transforming, and aggregating the data using tools like Apache Spark or Hadoop.
Data storage: Storing the processed data in data warehouses or data lakes like Amazon S3 ...
Use Spark (Databricks) notebooks to migrate 1000s of tables efficiently.
Utilize Spark's parallel processing capabilities to handle large volumes of data
Leverage Databricks notebooks for interactive data exploration and transformation
Automate the migration process using scripts or workflows
Optimize performance by tuning Spark configurations and cluster settings
Seeking new challenges and opportunities for growth in a more dynamic environment.
Desire for new challenges and learning opportunities
Seeking a more dynamic work environment
Looking to expand skill set and experience
Interested in working on cutting-edge technologies
Seeking better career advancement prospects
I expect EPAM to provide challenging projects, opportunities for growth, a collaborative work environment, and support for continuous learning.
Challenging projects that allow me to utilize my skills and knowledge
Opportunities for professional growth and advancement within the company
A collaborative work environment where teamwork is valued
Support for continuous learning through training programs and resources
Yes, I am willing to relocate for the right opportunity. I can join the company within 4 weeks.
Willing to relocate for the right opportunity
Can join within 4 weeks
Open to discussing relocation assistance if needed
What people are saying about EPAM Systems
I was interviewed in Feb 2025.
EPAM Systems interview questions for popular designations
I applied via Naukri.com and was interviewed in Nov 2024. There was 1 interview round.
Best practices for writing PowerShell scripts
Use clear and descriptive variable names
Comment your code to explain complex logic
Handle errors gracefully with try/catch blocks
Use functions to modularize your script
Avoid hardcoding values whenever possible
To create a Private Endpoint for Azure Key Vaults, you need to configure a private link service.
Navigate to the Azure portal and search for 'Key Vaults'.
Select the Key Vault you want to create a Private Endpoint for.
In the Key Vault settings, go to 'Private endpoint connections' and click on 'Add'.
Choose the subscription, resource group, and private DNS zone for the Private Endpoint.
Review and create the Private Endpoi
AKS allows for network configuration during creation and provides options for managing it.
During creation of AKS, network configuration options include specifying virtual network, subnet, and network policies.
Network configuration can be managed through Azure portal, Azure CLI, or ARM templates.
AKS supports network policies like Azure CNI, Kubenet, and Calico for network security and isolation.
Network configuration can...
Get interview-ready with Top EPAM Systems Interview Questions
I was interviewed in Feb 2025.
Questions on Python: 5 different simple python coding questions
I was interviewed in Jan 2025.
Stream based problems involve processing data in a continuous flow rather than all at once.
Use stream processing libraries like Apache Kafka or Apache Flink
Consider factors like data volume, velocity, and variety
Implement backpressure mechanisms to handle high data loads
I applied via LinkedIn and was interviewed in Dec 2024. There were 2 interview rounds.
Optional in Swift allows variables to have no value. Optional binding and chaining are used to safely unwrap optionals.
Optional in Swift allows variables to have no value, denoted by a '?' after the type.
Optional binding is used to safely unwrap optionals by checking if they contain a value.
Optional chaining allows you to call methods, access properties, and subscript on an optional that might currently be nil.
Example:...
ARC stands for Automatic Reference Counting, a memory management system used in iOS to automatically manage memory allocation and deallocation.
ARC automatically tracks and manages the memory used by objects in an iOS application.
It keeps track of the number of references to an object and deallocates the object when there are no more references to it.
ARC is enabled by default in Xcode for iOS projects, reducing the need...
Closures are self-contained blocks of functionality that can be passed around and used in code.
Closures capture and store references to any constants and variables from the context in which they are defined.
To prevent strong reference cycles, use capture lists in closures.
Use weak or unowned references when capturing self inside a closure to avoid memory leaks.
Middleware is software that acts as a bridge between an operating system or database and applications, allowing them to communicate with each other.
Middleware is a layer of software that sits between the operating system and applications, providing services such as authentication, logging, and caching.
Custom middleware can be created in ASP.NET Core by implementing the IMiddleware interface and adding it to the applica...
Yes, a program can be written to convert 'aabbccaaa' to '2a2b2c3a'.
Create a function that iterates through the input string and counts the consecutive characters.
Store the count and character in a new string as needed.
Return the final output string.
I applied via Approached by Company and was interviewed in Oct 2024. There were 3 interview rounds.
TestNG annotations allow for parallel execution of test methods using attributes like 'parallel' and 'thread-count'.
TestNG annotations like @Test, @BeforeTest, @AfterTest can be used with attributes like 'parallel' to specify parallel execution of test methods.
The 'parallel' attribute can have values like 'methods', 'tests', 'classes', 'instances' to define the scope of parallel execution.
The 'thread-count' attribute c...
I was interviewed in Jan 2025.
Some of the top questions asked at the EPAM Systems interview -
The duration of EPAM Systems interview process can vary, but typically it takes about less than 2 weeks to complete.
based on 488 interviews
Interview experience
based on 1.4k reviews
Rating in categories
Senior Software Engineer
2.8k
salaries
| ₹0 L/yr - ₹0 L/yr |
Software Engineer
1.8k
salaries
| ₹0 L/yr - ₹0 L/yr |
Lead Software Engineer
871
salaries
| ₹0 L/yr - ₹0 L/yr |
Senior Systems Engineer
305
salaries
| ₹0 L/yr - ₹0 L/yr |
Software Test Automation Engineer
264
salaries
| ₹0 L/yr - ₹0 L/yr |
TCS
Infosys
Wipro
HCLTech