Upload Button Icon Add office photos

Filter interviews by

Sisense Data Engineer Interview Questions and Answers

Updated 20 Jun 2024

Sisense Data Engineer Interview Experiences

1 interview found

Data Engineer Interview Questions & Answers

user image Anonymous

posted on 20 Jun 2024

Interview experience
3
Average
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Selected Selected

I applied via LinkedIn and was interviewed before Jun 2023. There was 1 interview round.

Round 1 - HR 

(2 Questions)

  • Q1. What made you choose Sisense?
  • Q2. Tell me about your experience working at a SaaS company?

Interview questions from similar companies

Interview experience
5
Excellent
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Technical 

(1 Question)

  • Q1. Sql commands group by and order by
Round 2 - Technical 

(1 Question)

  • Q1. Cicd pipeline and kafka
Round 3 - Client Interview 

(1 Question)

  • Q1. Reparitition and pipelines
Round 4 - Technical 

(1 Question)

  • Q1. Joins operations
Interview experience
1
Bad
Difficulty level
Moderate
Process Duration
6-8 weeks
Result
Not Selected

I was interviewed in Dec 2024.

Round 1 - Technical 

(2 Questions)

  • Q1. They asked me regarding project
  • Q2. Sql queries and data bricks related questions
Round 2 - Technical 

(1 Question)

  • Q1. Details explaination on project that I have worked and optimization techniques

Interview Preparation Tips

Interview preparation tips for other job seekers - I have successfully cleared the L1 and L2 rounds, and my background verification has also been completed. I was informed that an offer would be sent the next day. However, the HR then stated that there would be a discussion solely about the CTC, with no technical questions involved. I agreed to this; yet, the interviewer proceeded to ask high-level technical questions that I had not worked on, leading to the final decision of not selecting me. It felt like a fake interview process, accompanied by delayed communication and a lack of response, resulting in a very negative experience with Tech Mahindra.
Interview experience
5
Excellent
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Not Selected

I applied via Naukri.com and was interviewed in Dec 2024. There was 1 interview round.

Round 1 - Technical 

(5 Questions)

  • Q1. Scenario based questions on Azure data factory and pipelines
  • Q2. Optimisation technic to improve the performance of databricks
  • Q3. What is Autoloader
  • Q4. What is unity catalog
  • Q5. How you do the alerting mechanism in adf for failed pipelines
Interview experience
3
Average
Difficulty level
Easy
Process Duration
Less than 2 weeks
Result
Not Selected

I applied via Company Website and was interviewed in Nov 2024. There were 2 interview rounds.

Round 1 - Technical 

(2 Questions)

  • Q1. Sql constrainsts, star schema, dml dcl commands
  • Q2. About cureent project and responsibilities
Round 2 - Technical 

(2 Questions)

  • Q1. Current projects and resposibilities
  • Q2. Where vs having, reason for job change

Interview Preparation Tips

Interview preparation tips for other job seekers - 1. Technical - about you current project and responsibilities, basic SQL question-constraints, starschema, DML DCL command, one sql query write.
2. Technical with senior manager- about project ,where vs having , reason of job change
Interview experience
3
Average
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
No response

I applied via Naukri.com and was interviewed in Oct 2024. There were 2 interview rounds.

Round 1 - Technical 

(7 Questions)

  • Q1. How do you optimize SQL queries?
  • Ans. 

    Optimizing SQL queries involves using indexes, avoiding unnecessary joins, and optimizing the query structure.

    • Use indexes on columns frequently used in WHERE clauses

    • Avoid using SELECT * and only retrieve necessary columns

    • Optimize joins by using INNER JOIN instead of OUTER JOIN when possible

    • Use EXPLAIN to analyze query performance and make necessary adjustments

  • Answered by AI
  • Q2. How do you do performance optimization in Spark. Tell how you did it in you project.
  • Ans. 

    Performance optimization in Spark involves tuning configurations, optimizing code, and utilizing caching.

    • Tune Spark configurations such as executor memory, number of executors, and shuffle partitions.

    • Optimize code by reducing unnecessary shuffles, using efficient transformations, and avoiding unnecessary data movements.

    • Utilize caching to store intermediate results in memory and avoid recomputation.

    • Example: In my projec...

  • Answered by AI
  • Q3. What is SparkContext and SparkSession?
  • Ans. 

    SparkContext is the main entry point for Spark functionality, while SparkSession is the entry point for Spark SQL.

    • SparkContext is the entry point for low-level API functionality in Spark.

    • SparkSession is the entry point for Spark SQL functionality.

    • SparkContext is used to create RDDs (Resilient Distributed Datasets) in Spark.

    • SparkSession provides a unified entry point for reading data from various sources and performing

  • Answered by AI
  • Q4. When a spark job is submitted, what happens at backend. Explain the flow.
  • Ans. 

    When a spark job is submitted, various steps are executed at the backend to process the job.

    • The job is submitted to the Spark driver program.

    • The driver program communicates with the cluster manager to request resources.

    • The cluster manager allocates resources (CPU, memory) to the job.

    • The driver program creates DAG (Directed Acyclic Graph) of the job stages and tasks.

    • Tasks are then scheduled and executed on worker nodes ...

  • Answered by AI
  • Q5. Calculate second highest salary using SQL as well as pyspark.
  • Ans. 

    Calculate second highest salary using SQL and pyspark

    • Use SQL query with ORDER BY and LIMIT to get the second highest salary

    • In pyspark, use orderBy() and take() functions to achieve the same result

  • Answered by AI
  • Q6. 2 types of modes for Spark architecture ?
  • Ans. 

    The two types of modes for Spark architecture are standalone mode and cluster mode.

    • Standalone mode: Spark runs on a single machine with a single JVM and is suitable for development and testing.

    • Cluster mode: Spark runs on a cluster of machines managed by a cluster manager like YARN or Mesos for production workloads.

  • Answered by AI
  • Q7. If you want very less latency - which is better standalone or client mode?
  • Ans. 

    Client mode is better for very less latency due to direct communication with the cluster.

    • Client mode allows direct communication with the cluster, reducing latency.

    • Standalone mode requires an additional layer of communication, increasing latency.

    • Client mode is preferred for real-time applications where low latency is crucial.

  • Answered by AI
Round 2 - Technical 

(2 Questions)

  • Q1. Scenario based. Write SQL and pyspark code for a dataset.
  • Q2. If you have to find latest record based on latest timestamp in a table for a particular customer(table is having history) , how will you do it. Self join and nested query will be expensive. Optimized query...

Interview Preparation Tips

Topics to prepare for LTIMindtree Data Engineer interview:
  • SQL
  • pyspark
  • ETL
Interview preparation tips for other job seekers - L2 was scheduled next day to L1 so the process is fast. Brush up your practical knowledge more.

Skills evaluated in this interview

Data Engineer Interview Questions & Answers

Genpact user image Sashikanta Parida

posted on 17 Dec 2024

Interview experience
5
Excellent
Difficulty level
Moderate
Process Duration
Less than 2 weeks
Result
Not Selected

I applied via Recruitment Consulltant and was interviewed in Nov 2024. There were 2 interview rounds.

Round 1 - Technical 

(3 Questions)

  • Q1. What are different type of joins available in Databricks?
  • Ans. 

    Different types of joins available in Databricks include inner join, outer join, left join, right join, and cross join.

    • Inner join: Returns only the rows that have matching values in both tables.

    • Outer join: Returns all rows when there is a match in either table.

    • Left join: Returns all rows from the left table and the matched rows from the right table.

    • Right join: Returns all rows from the right table and the matched rows ...

  • Answered by AI
  • Q2. How do you make your data pipeline fault tolerant?
  • Ans. 

    Implementing fault tolerance in a data pipeline involves redundancy, monitoring, and error handling.

    • Use redundant components to ensure continuous data flow

    • Implement monitoring tools to detect failures and bottlenecks

    • Set up automated alerts for immediate response to issues

    • Design error handling mechanisms to gracefully handle failures

    • Use checkpoints and retries to ensure data integrity

  • Answered by AI
  • Q3. What is AutoLoader?
  • Ans. 

    AutoLoader is a feature in data engineering that automatically loads data from various sources into a data warehouse or database.

    • Automates the process of loading data from different sources

    • Reduces manual effort and human error

    • Can be scheduled to run at specific intervals

    • Examples: Apache Nifi, AWS Glue

  • Answered by AI
Round 2 - Technical 

(2 Questions)

  • Q1. How do you connect to different services in Azure?
  • Ans. 

    To connect to different services in Azure, you can use Azure SDKs, REST APIs, Azure Portal, Azure CLI, and Azure PowerShell.

    • Use Azure SDKs for programming languages like Python, Java, C#, etc.

    • Utilize REST APIs to interact with Azure services programmatically.

    • Access and manage services through the Azure Portal.

    • Leverage Azure CLI for command-line interface interactions.

    • Automate tasks using Azure PowerShell scripts.

  • Answered by AI
  • Q2. What are linked Services?
  • Ans. 

    Linked Services are connections to external data sources or destinations in Azure Data Factory.

    • Linked Services define the connection information needed to connect to external data sources or destinations.

    • They can be used in Data Factory pipelines to read from or write to external systems.

    • Examples of Linked Services include Azure Blob Storage, Azure SQL Database, and Amazon S3.

  • Answered by AI

Data Engineer Interview Questions & Answers

Wipro user image Lakshmi Narayana

posted on 27 Nov 2024

Interview experience
4
Good
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Technical 

(2 Questions)

  • Q1. Explain adf questions in detail
  • Ans. 

    ADF questions refer to Azure Data Factory questions which are related to data integration and data transformation processes.

    • ADF questions are related to Azure Data Factory, a cloud-based data integration service.

    • These questions may involve data pipelines, data flows, activities, triggers, and data movement.

    • Candidates may be asked about their experience with designing, monitoring, and managing data pipelines in ADF.

    • Exam...

  • Answered by AI
  • Q2. Project related questions
Round 2 - Technical 

(2 Questions)

  • Q1. Project data related questions
  • Q2. Databricks and SQL interview questions
Interview experience
4
Good
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Technical 

(3 Questions)

  • Q1. What are the optimization techniques used in Apache Spark?
  • Q2. 2 SQL queries , 1 PySpark code and 1 Python Code .
  • Q3. 2-3 Scenario Based questions from ADF and databricks .
Interview experience
2
Poor
Difficulty level
-
Process Duration
-
Result
-
Round 1 - Coding Test 

Oops dsa sql network

Round 2 - Technical 

(2 Questions)

  • Q1. Dsa based python question
  • Q2. Dsa based python question on tress
Round 3 - Technical 

(2 Questions)

  • Q1. Sql question queries
  • Q2. Sql question query

Sisense Interview FAQs

How many rounds are there in Sisense Data Engineer interview?
Sisense interview process usually has 1 rounds. The most common rounds in the Sisense interview process are HR.

Tell us how to improve this page.

Data Engineer Interview Questions from Similar Companies

View all
Tester
3 salaries
unlock blur

₹5.2 L/yr - ₹9.2 L/yr

Explore more salaries
Compare Sisense with

TCS

3.7
Compare

Accenture

3.9
Compare

Wipro

3.7
Compare

Cognizant

3.8
Compare
Did you find this page helpful?
Yes No
write
Share an Interview