Filter interviews by
I applied via LinkedIn and was interviewed in Nov 2024. There were 4 interview rounds.
My current day-to-day tasks involve designing and implementing data pipelines, optimizing data storage and retrieval, and collaborating with cross-functional teams.
Designing and implementing data pipelines to extract, transform, and load data from various sources
Optimizing data storage and retrieval processes for efficiency and scalability
Collaborating with cross-functional teams to understand data requirements and del...
The end-to-end project architecture involves designing and implementing the entire data pipeline from data ingestion to data visualization.
Data ingestion: Collecting data from various sources such as databases, APIs, and files.
Data processing: Cleaning, transforming, and aggregating the data using tools like Apache Spark or Hadoop.
Data storage: Storing the processed data in data warehouses or data lakes like Amazon S3 ...
Use Spark (Databricks) notebooks to migrate 1000s of tables efficiently.
Utilize Spark's parallel processing capabilities to handle large volumes of data
Leverage Databricks notebooks for interactive data exploration and transformation
Automate the migration process using scripts or workflows
Optimize performance by tuning Spark configurations and cluster settings
Seeking new challenges and opportunities for growth in a more dynamic environment.
Desire for new challenges and learning opportunities
Seeking a more dynamic work environment
Looking to expand skill set and experience
Interested in working on cutting-edge technologies
Seeking better career advancement prospects
I expect EPAM to provide challenging projects, opportunities for growth, a collaborative work environment, and support for continuous learning.
Challenging projects that allow me to utilize my skills and knowledge
Opportunities for professional growth and advancement within the company
A collaborative work environment where teamwork is valued
Support for continuous learning through training programs and resources
Yes, I am willing to relocate for the right opportunity. I can join the company within 4 weeks.
Willing to relocate for the right opportunity
Can join within 4 weeks
Open to discussing relocation assistance if needed
I applied via Naukri.com and was interviewed in Nov 2024. There were 2 interview rounds.
Big data related multiple questions around 10 Multiple choice
One sql question to find the salary details medium difficulty
One python related question
What people are saying about EPAM Systems
I applied via Approached by Company and was interviewed in Dec 2023. There were 4 interview rounds.
The question is asking about types of transformations, number of jobs, tasks, and actions in the context of a Senior Data Engineer role.
Types of transformations: Extract, Transform, Load (ETL), MapReduce, Spark transformations, SQL transformations
Number of jobs: Depends on the complexity and scale of the data engineering projects
Number of tasks: Varies based on the number of data sources, data transformations, and data...
Spark is a distributed processing engine, Airflow is a workflow management system, and BigQuery is a fully managed data warehouse.
Spark is designed for big data processing and provides in-memory computation capabilities.
Airflow is used for orchestrating and scheduling data pipelines.
BigQuery is a serverless data warehouse that allows for fast and scalable analytics.
Spark can be integrated with Airflow to schedule and m...
Optimization techniques in Spark, SQL, BigQuery, and Airflow.
Use partitioning and bucketing in Spark to optimize data processing.
Optimize SQL queries by using indexes, query rewriting, and query optimization techniques.
In BigQuery, use partitioning and clustering to improve query performance.
Leverage Airflow's task parallelism and resource allocation to optimize workflow execution.
To delete duplicates from a table in Spark and SQL, you can use the DISTINCT keyword or the dropDuplicates() function.
In SQL, you can use the DISTINCT keyword in a SELECT statement to retrieve unique rows from a table.
In Spark, you can use the dropDuplicates() function on a DataFrame to remove duplicate rows.
Both methods compare all columns by default, but you can specify specific columns to consider for duplicates.
You...
Dataflow and Dataproc are both processing services in GCP, but with different approaches and use cases.
Dataflow is a fully managed service for executing batch and streaming data processing pipelines.
Dataproc is a managed Spark and Hadoop service for running big data processing and analytics workloads.
Dataflow provides a serverless and auto-scaling environment, while Dataproc offers more control and flexibility.
Dataflow...
This was final round with Client.
They ask questions based on my work in pyspark area.
Questions are like:
What kind of transformations you used.
Broadcast join internals.
Spark internal joins
Spark catalyst optimizer.What are the joins happens in catalyst optimizer.
Window function question: 3rd highest salary of an emp
Discussion on airflow arch and how to deploy a airflow dag in gcp.
Discussion on BQ on what kind of work i have done till now.
EPAM Systems interview questions for designations
I applied via LinkedIn and was interviewed in Mar 2024. There were 3 interview rounds.
Get interview-ready with Top EPAM Systems Interview Questions
I applied via Naukri.com and was interviewed in Sep 2024. There was 1 interview round.
Python questions were asked in this round
I applied via Naukri.com and was interviewed in Mar 2023. There were 5 interview rounds.
2 coding question in Python where 1 of them is basic level and other one is based on binary tree. 3 question multiple choice question consists of 10 question which will be based on basic of spark and big data concepts.
Python
Sql coding tests
I applied via Approached by Company and was interviewed before Mar 2023. There were 2 interview rounds.
I applied via Recruitment Consulltant and was interviewed before Dec 2021. There were 4 interview rounds.
2 questions from python and one from SQL average questions from hacker earth
2 Interview rounds
based on 7 reviews
Rating in categories
Senior Software Engineer
2.6k
salaries
| ₹15 L/yr - ₹42.8 L/yr |
Software Engineer
1.7k
salaries
| ₹6.9 L/yr - ₹24 L/yr |
Lead Software Engineer
831
salaries
| ₹18 L/yr - ₹52 L/yr |
Senior Systems Engineer
309
salaries
| ₹12 L/yr - ₹38.8 L/yr |
Software Test Automation Engineer
265
salaries
| ₹6 L/yr - ₹20 L/yr |
TCS
Infosys
Wipro
HCLTech