i
Innominds Software
Filter interviews by
I applied via Campus Placement and was interviewed in Apr 2022. There were 3 interview rounds.
Python questions
Interview questions for Big Data Engineer role
Java and Python are both popular programming languages for Big Data processing, but Java is preferred for its performance and scalability
Normalization is the process of organizing data in a database to reduce redundancy and improve data integrity
MongoDB is a NoSQL database that is highly scalable and flexible, making it a good choice for Big Data applications
To reverse a li...
Top trending discussions
I applied via AmbitionBox and was interviewed in Nov 2024. There were 4 interview rounds.
I utilize tools such as Excel, Python, SQL, and Tableau for data analysis.
Excel for basic data manipulation and visualization
Python for advanced data analysis and machine learning
SQL for querying databases
Tableau for creating interactive visualizations
Data analysis of code in the context of data analysis.
Coding logical question paper.
posted on 28 Aug 2024
I have experience working on projects involving data pipeline development, ETL processes, and data warehousing.
Developed ETL processes to extract, transform, and load data from various sources into a data warehouse
Built data pipelines to automate the flow of data between systems and ensure data quality and consistency
Optimized database performance and implemented data modeling best practices
Worked on real-time data pro...
Distribution in Spark refers to how data is divided across different nodes in a cluster for parallel processing.
Data is partitioned across multiple nodes in a cluster to enable parallel processing
Distribution can be controlled using partitioning techniques like hash partitioning or range partitioning
Ensures efficient utilization of resources and faster processing times
AWS Glue can process petabytes of data per hour
AWS Glue can process petabytes of data per hour, depending on the configuration and resources allocated
It is designed to scale horizontally to handle large volumes of data efficiently
AWS Glue can be used for ETL (Extract, Transform, Load) processes on massive datasets
Distribution in Spark refers to how data is divided across different nodes in a cluster for parallel processing.
Distribution in Spark determines how data is partitioned across different nodes in a cluster
It helps in achieving parallel processing by distributing the workload
Examples of distribution methods in Spark include hash partitioning and range partitioning
AWS Glue can process petabytes of data per hour.
AWS Glue can process petabytes of data per hour, making it suitable for large-scale data processing tasks.
It can handle various types of data sources, including structured and semi-structured data.
AWS Glue offers serverless ETL (Extract, Transform, Load) capabilities, allowing for scalable and cost-effective data processing.
It integrates seamlessly with other AWS services...
Spark is a fast and general-purpose cluster computing system, while PySpark is the Python API for Spark.
Spark is a distributed computing system that provides an interface for programming entire clusters with implicit data parallelism and fault tolerance.
PySpark is the Python API for Spark that allows developers to write Spark applications using Python.
Spark and PySpark are commonly used for big data processing, machine...
I applied via Company Website and was interviewed in Jul 2024. There was 1 interview round.
Pods are the smallest deployable units in Kubernetes, consisting of one or more containers.
Pods are used to run and manage containers in Kubernetes
Each pod has its own unique IP address within the Kubernetes cluster
Pods can contain multiple containers that share resources and are scheduled together
Pods are ephemeral and can be easily created, destroyed, or replicated
Pods can be managed and scaled using Kubernetes contr
A stored procedure is a set of SQL statements that can be saved and reused in a database.
Stored procedures can improve performance by reducing network traffic and improving security.
They can be used to encapsulate business logic and complex queries.
Stored procedures can accept input parameters and return output parameters or result sets.
I applied via Approached by Company and was interviewed in May 2024. There was 1 interview round.
To create a DBT project, you need to set up a project directory, create models, define sources, and run tests.
Set up a project directory with a dbt_project.yml file
Create models in the models directory using SQL files
Define sources in the sources.yml file
Run tests using dbt test command
Materializations in dbt are pre-computed tables that store the results of dbt models for faster query performance.
Materializations are created using the 'materialized' parameter in dbt models.
Common types of materializations include 'view', 'table', and 'incremental'.
Materializations help improve query performance by reducing the need to recompute data on every query.
Materializations can be refreshed manually or automa
dbt snapshots are a way to capture the state of your data model at a specific point in time.
dbt snapshots are used to create point-in-time snapshots of your data model
They allow you to track changes in your data over time
Snapshots can be used for auditing, debugging, or creating historical reports
I have created interactive dashboards using Tableau to visualize and analyze data for various projects.
Utilized Tableau to connect to data sources and create interactive visualizations
Designed dashboards with filters, drill-down capabilities, and dynamic elements
Included key performance indicators (KPIs) and trend analysis in the dashboards
Used color coding and data labels to enhance data interpretation
Shared dashboard
I applied via Approached by Company and was interviewed in Mar 2024. There were 2 interview rounds.
Coding test for all the student attending the first round, give your best practice every day all the best in your interview
My interests include data analysis, problem-solving, and continuous learning.
Data analysis
Problem-solving
Continuous learning
Senior Software Engineer
447
salaries
| ₹7.5 L/yr - ₹27 L/yr |
Software Engineer
422
salaries
| ₹4 L/yr - ₹14 L/yr |
Senior Engineer
210
salaries
| ₹6.2 L/yr - ₹25.1 L/yr |
Associate Software Engineer
166
salaries
| ₹2.2 L/yr - ₹8.5 L/yr |
Engineer
158
salaries
| ₹3 L/yr - ₹13.2 L/yr |
Persistent Systems
LTIMindtree
Mphasis
TCS