i
IBM
Proud winner of ABECA 2024 - AmbitionBox Employee Choice Awards
Filter interviews by
I applied via Approached by Company and was interviewed in Jun 2024. There were 2 interview rounds.
Python coding question and couple of SQL questions
Spark optimization techniques focus on improving performance and efficiency of Spark jobs.
Partitioning data to optimize parallelism
Caching frequently accessed data
Using broadcast variables for small lookup tables
Avoiding shuffling operations whenever possible
Tuning memory settings for optimal performance
I have faced difficulties in handling large volumes of data, ensuring data quality, and managing dependencies in ETL pipelines.
Handling large volumes of data can lead to performance issues and scalability challenges.
Ensuring data quality involves dealing with data inconsistencies, errors, and missing values.
Managing dependencies between different stages of the ETL process can be complex and prone to failures.
I was interviewed in Aug 2024.
Python and sql tasks
What people are saying about IBM
Bigquery Architecture, Project Discussion
Python and SQL questions
IBM interview questions for designations
I applied via Naukri.com and was interviewed in Oct 2022. There were 2 interview rounds.
Questions on big data, Hadoop, Spark, Scala, Git, project and Agile.
Hadoop architecture and HDFS commands for copying and listing files in HDFS
Spark architecture and Transformation and Action question
What happens when we submit a Spark program
Spark DataFrame coding question
Scala basic program on List
Git and Github
Project-related question
Agile-related
I applied via Naukri.com and was interviewed in May 2022. There was 1 interview round.
I was asked to do some python code provided by the interviewer and some senario-based SQL queries and a lot of job processing theory and optimization techniques used in my project.
Optimization techniques used in project
Caching
Parallel processing
Compression
Indexing
Query optimization
I applied via Recruitment Consulltant
I applied via Naukri.com and was interviewed in Nov 2024. There was 1 interview round.
I am a Senior Data Engineer with experience in building scalable data pipelines and optimizing data processing workflows.
Experience in designing and implementing ETL processes using tools like Apache Spark and Airflow
Proficient in working with large datasets and optimizing query performance
Strong background in data modeling and database design
Worked on projects involving real-time data processing and streaming analytic
Decorators in Python are functions that modify the behavior of other functions or methods.
Decorators are defined using the @decorator_name syntax before a function definition.
They can be used to add functionality to existing functions without modifying their code.
Decorators can be used for logging, timing, authentication, and more.
Example: @staticmethod decorator in Python is used to define a static method in a class.
SQL query to group by employee ID and combine first name and last name with a space
Use the GROUP BY clause to group by employee ID
Use the CONCAT function to combine first name and last name with a space
Select employee ID, CONCAT(first_name, ' ', last_name) AS full_name
Constructors in Python are special methods used for initializing objects. They are called automatically when a new instance of a class is created.
Constructors are defined using the __init__() method in a class.
They are used to initialize instance variables of a class.
Example: class Person: def __init__(self, name, age): self.name = name self.age = age person1 = Person('Alice', 30)
Indexing in SQL is a technique used to improve the performance of queries by creating a data structure that allows for faster retrieval of data.
Indexes are created on columns in a database table to speed up the retrieval of rows that match a certain condition in a WHERE clause.
Indexes can be created using CREATE INDEX statement in SQL.
Types of indexes include clustered indexes, non-clustered indexes, unique indexes, an...
Spark works well with Parquet files due to its columnar storage format, efficient compression, and ability to push down filters.
Parquet files are columnar storage format, which aligns well with Spark's processing model of working on columns rather than rows.
Parquet files support efficient compression, reducing storage space and improving read performance in Spark.
Spark can push down filters to Parquet files, allowing f...
I applied via Naukri.com and was interviewed in Nov 2024. There were 2 interview rounds.
based on 7 interviews
2 Interview rounds
based on 51 reviews
Rating in categories
Application Developer
11.7k
salaries
| ₹5.5 L/yr - ₹24 L/yr |
Software Engineer
5.5k
salaries
| ₹5.5 L/yr - ₹22.5 L/yr |
Advisory System Analyst
5.2k
salaries
| ₹9.4 L/yr - ₹29.8 L/yr |
Senior Software Engineer
4.8k
salaries
| ₹8 L/yr - ₹30 L/yr |
Senior Systems Engineer
4.5k
salaries
| ₹5.7 L/yr - ₹20.8 L/yr |
Oracle
TCS
Cognizant
Accenture