Upload Button Icon Add office photos

Filter interviews by

EPAM Systems Big Data Engineer Interview Questions and Answers

Updated 24 Jul 2024

EPAM Systems Big Data Engineer Interview Experiences

2 interviews found

Interview experience
1
Bad
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Technical 

(1 Question)

  • Q1. About last worked project?

I applied via Recruitment Consulltant and was interviewed before Mar 2021. There was 1 interview round.

Round 1 - Coding Test 

Java programming questions in hanker rank

Interview Preparation Tips

Interview preparation tips for other job seekers - Spark interview questions and AWS if you have worked on

Big Data Engineer Interview Questions Asked at Other Companies

Q1. Difference between partitioning and bucketing. Types of joins in ... read more
asked in TCS
Q2. What optimization techniques have you utilized in your projects? ... read more
Q3. Checking whether a fibonacci number is present between a particuk ... read more
asked in TCS
Q4. What is the difference between lineage and directed acyclic graph ... read more
asked in TCS
Q5. What is the difference between cache and persistence?

Interview questions from similar companies

Interview experience
1
Bad
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Not Selected

I applied via Naukri.com and was interviewed in Jul 2023. There were 2 interview rounds.

Round 1 - Resume Shortlist 
Pro Tip by AmbitionBox:
Keep your resume crisp and to the point. A recruiter looks at your resume for an average of 6 seconds, make sure to leave the best impression.
View all tips
Round 2 - Technical 

(2 Questions)

  • Q1. Basic Questions of Scala Functional Programming concepts.
  • Q2. Spark internal working and optimization techniques
  • Ans. 

    Spark internal working and optimization techniques

    • Spark uses Directed Acyclic Graph (DAG) for optimizing workflows

    • Lazy evaluation helps in optimizing transformations by combining them into a single stage

    • Caching and persistence of intermediate results can improve performance

    • Partitioning data can help in parallel processing and reducing shuffle operations

  • Answered by AI

Interview Preparation Tips

Interview preparation tips for other job seekers - The interview call was abruptly terminated into 20 mins of call duration, as the HR had another conflicting call. The HR called me over cellphone and told that if the interview panel requested she will let me know and the call can be extended, but the HR did not call, The interview did not extended. Finally they rejected me just after the panel spoke 20 mins with me in the 1st round interview. This shows how unprofessional are they in scheduling an interview call and how could any panel can decide within 20 mins of a discussion. Definitely not recommending anyone to attend Bigdata Engineering interviews here.

Skills evaluated in this interview

Interview Questionnaire 

2 Questions

  • Q1. Currently Working Project details, Cassandra and hive difference , How to decide no of executers in cluster config
  • Q2. Spark and hadoop architectural difference , DAG, What is stage boundaries , Partitioning and bucketing in hive
  • Ans. 

    Spark and Hadoop have different architectures. DAG is a directed acyclic graph. Stage boundaries are logical divisions in a Spark job. Hive has partitioning and bucketing.

    • Spark is an in-memory processing engine while Hadoop is a distributed file system.

    • DAG is a graph of stages in a Spark job.

    • Stage boundaries are logical divisions in a Spark job that help optimize execution.

    • Partitioning in Hive is a way to divide a tabl...

  • Answered by AI

Skills evaluated in this interview

Interview experience
5
Excellent
Difficulty level
Moderate
Process Duration
6-8 weeks
Result
Selected Selected
Round 1 - Technical 

(1 Question)

  • Q1. More on Technical area
Round 2 - Technical 

(1 Question)

  • Q1. More on Technical area
Round 3 - One-on-one 

(1 Question)

  • Q1. Technical + Behaviour
Round 4 - One-on-one 

(1 Question)

  • Q1. Technical + Behaviour
Round 5 - HR 

(1 Question)

  • Q1. Expectation and Genaral
Interview experience
3
Average
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Not Selected

I applied via Naukri.com and was interviewed in Nov 2024. There was 1 interview round.

Round 1 - Technical 

(2 Questions)

  • Q1. How do you utilize the enhanced optimization option in AWS Glue?
  • Ans. 

    Enhanced optimization in AWS Glue improves job performance by automatically adjusting resources based on workload

    • Enhanced optimization in AWS Glue automatically adjusts resources like DPUs based on workload

    • It helps improve job performance by optimizing resource allocation

    • Users can enable enhanced optimization in AWS Glue job settings

  • Answered by AI
  • Q2. What are the best practices for optimizing querying in Amazon Redshift?
  • Ans. 

    Optimizing querying in Amazon Redshift involves proper table design, distribution keys, sort keys, and query optimization techniques.

    • Use appropriate distribution keys to evenly distribute data across nodes for parallel processing.

    • Utilize sort keys to physically order data on disk, reducing the need for sorting during queries.

    • Avoid using SELECT * and instead specify only the columns needed to reduce data transfer.

    • Use AN...

  • Answered by AI
Interview experience
1
Bad
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Technical 

(1 Question)

  • Q1. Data engineer roles and resposibilities

Interview Preparation Tips

Interview preparation tips for other job seekers - Don't Go, worst management service and lots of office politics.
Interview experience
5
Excellent
Difficulty level
Hard
Process Duration
2-4 weeks
Result
-

I applied via Naukri.com and was interviewed in Sep 2024. There was 1 interview round.

Round 1 - One-on-one 

(2 Questions)

  • Q1. What is scd type 2?
  • Ans. 

    SCD type 2 is a method used in data warehousing to track historical changes by creating a new record for each change.

    • SCD type 2 stands for Slowly Changing Dimension type 2

    • It involves creating a new record in the dimension table whenever there is a change in the data

    • The old record is marked as inactive and the new record is marked as current

    • It allows for historical tracking of changes in data over time

    • Example: If a cust...

  • Answered by AI
  • Q2. Pyspark question read CSV from folder and add column in each csv file and write it to different location.

Skills evaluated in this interview

Interview experience
4
Good
Difficulty level
Hard
Process Duration
-
Result
-

I applied via Campus Placement and was interviewed in Jul 2024. There was 1 interview round.

Round 1 - Technical 

(2 Questions)

  • Q1. Questions about cables and linux
  • Q2. Questions about networking and amazon principles
Interview experience
3
Average
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Not Selected

I applied via Job Portal and was interviewed in Jul 2024. There was 1 interview round.

Round 1 - HR 

(2 Questions)

  • Q1. Tell about experience
  • Ans. 

    I have over 5 years of experience in data engineering, working with large datasets and implementing data pipelines.

    • Developed and maintained ETL processes to extract, transform, and load data from various sources

    • Optimized database performance and implemented data quality checks

    • Worked with cross-functional teams to design and implement data solutions

    • Utilized tools such as Apache Spark, Hadoop, and SQL for data processing

    • ...

  • Answered by AI
  • Q2. What would you do when you are given a new task to do
  • Ans. 

    I would start by understanding the requirements, breaking down the task into smaller steps, researching if needed, and then creating a plan to execute the task efficiently.

    • Understand the requirements of the task

    • Break down the task into smaller steps

    • Research if needed to gather necessary information

    • Create a plan to execute the task efficiently

    • Communicate with stakeholders for clarification or updates

    • Regularly track prog

  • Answered by AI

EPAM Systems Interview FAQs

How many rounds are there in EPAM Systems Big Data Engineer interview?
EPAM Systems interview process usually has 1 rounds. The most common rounds in the EPAM Systems interview process are Coding Test and Technical.
How to prepare for EPAM Systems Big Data Engineer interview?
Go through your CV in detail and study all the technologies mentioned in your CV. Prepare at least two technologies or languages in depth if you are appearing for a technical interview at EPAM Systems. The most common topics and skills that interviewers at EPAM Systems expect are Big Data, Spark, GCP, Apache and RDBMS.

Tell us how to improve this page.

EPAM Systems Big Data Engineer Salary
based on 21 salaries
₹10 L/yr - ₹30 L/yr
96% more than the average Big Data Engineer Salary in India
View more details

EPAM Systems Big Data Engineer Reviews and Ratings

based on 1 review

2.0/5

Rating in categories

1.0

Skill development

4.0

Work-Life balance

3.0

Salary & Benefits

2.0

Job Security

2.0

Company culture

2.0

Promotions/Appraisal

3.0

Work Satisfaction

Explore 1 Review and Rating
Senior Software Engineer
2.6k salaries
unlock blur

₹15 L/yr - ₹42.7 L/yr

Software Engineer
1.7k salaries
unlock blur

₹6.9 L/yr - ₹24 L/yr

Lead Software Engineer
831 salaries
unlock blur

₹18 L/yr - ₹52 L/yr

Senior Systems Engineer
304 salaries
unlock blur

₹12 L/yr - ₹36.3 L/yr

Software Test Automation Engineer
267 salaries
unlock blur

₹7 L/yr - ₹20 L/yr

Explore more salaries
Compare EPAM Systems with

TCS

3.7
Compare

Infosys

3.7
Compare

Wipro

3.7
Compare

HCLTech

3.5
Compare

Calculate your in-hand salary

Confused about how your in-hand salary is calculated? Enter your annual salary (CTC) and get your in-hand salary
Did you find this page helpful?
Yes No
write
Share an Interview