i
Perficient
Filter interviews by
Tools used for data engineering include ETL tools, programming languages, databases, and cloud platforms.
ETL tools like Apache NiFi, Talend, and Informatica are used for data extraction, transformation, and loading.
Programming languages like Python, Java, and Scala are used for data processing and analysis.
Databases like MySQL, PostgreSQL, and MongoDB are used for storing and managing data.
Cloud platforms like AWS, Azu...
Blockers faced while working as a Data Engineer
Lack of proper documentation
Inadequate infrastructure
Data quality issues
Limited access to necessary data sources
Inefficient data processing pipelines
I applied via Campus Placement and was interviewed before May 2023. There was 1 interview round.
Top trending discussions
I was interviewed in Oct 2024.
Designing an ADF pipeline for data processing
Identify data sources and destinations
Define data transformations and processing steps
Consider scheduling and monitoring requirements
Utilize ADF activities like Copy Data, Data Flow, and Databricks
Implement error handling and logging mechanisms
Discussing expected and current salary for negotiation purposes.
Be honest about your current salary and provide a realistic expectation for your desired salary.
Highlight your skills and experience that justify your desired salary.
Be open to negotiation and willing to discuss other benefits besides salary.
Research industry standards and salary ranges for similar positions to support your negotiation.
Focus on the value y...
I applied via Naukri.com and was interviewed in Oct 2024. There were 2 interview rounds.
Spark performance problems can arise due to inefficient code, data skew, resource constraints, and improper configuration.
Inefficient code can lead to slow performance, such as using collect() on large datasets.
Data skew can cause uneven distribution of data across partitions, impacting processing time.
Resource constraints like insufficient memory or CPU can result in slow Spark jobs.
Improper configuration settings, su...
I applied via Campus Placement
Based on SQL , statistics , python , cognitive
Address toxic work culture by open communication, setting boundaries, seeking support, and considering leaving if necessary.
Open communication with colleagues and management about issues
Set boundaries to protect your mental and emotional well-being
Seek support from HR, a mentor, or a therapist if needed
Consider leaving the toxic work environment if the situation does not improve
I was interviewed in Aug 2024.
I applied via Naukri.com and was interviewed in Oct 2024. There was 1 interview round.
Incremental load in pyspark refers to loading only new or updated data into a dataset without reloading the entire dataset.
Use the 'delta' function in pyspark to perform incremental loads by specifying the 'mergeSchema' option.
Utilize the 'partitionBy' function to optimize incremental loads by partitioning the data based on specific columns.
Implement a logic to identify new or updated records based on timestamps or uni...
based on 2 interviews
Interview experience
based on 7 reviews
Rating in categories
Technical Consultant
840
salaries
| ₹4.8 L/yr - ₹14.9 L/yr |
Senior Technical Consultant
430
salaries
| ₹7.1 L/yr - ₹21 L/yr |
Associate Technical Consultant
320
salaries
| ₹3.3 L/yr - ₹8 L/yr |
Technical Leader Consultant
207
salaries
| ₹11 L/yr - ₹29 L/yr |
Softwaretest Engineer
187
salaries
| ₹3.2 L/yr - ₹8.1 L/yr |
Cognizant
TCS
Infosys
Wipro