i
HCLTech
Filter interviews by
I applied via Recruitment Consultant and was interviewed before Nov 2020. There was 1 interview round.
Different types of Lookup in ETL
Cached Lookup
Dynamic Lookup
Unconnected Lookup
Connected Lookup
Static Lookup
Difference between Row_num and rowid in SQL and removing duplicates
ROW_NUM is a function that assigns a unique number to each row in a result set
ROWID is a unique identifier for a row in a table
To remove duplicates, use the DISTINCT keyword in a SELECT statement
Another way to remove duplicates is to use the GROUP BY clause
The HAVING clause can be used to filter out duplicates based on a condition
What people are saying about HCLTech
Validation of data between source and target is a crucial step in ETL process.
Compare the data types and formats of source and target data
Check for missing or extra data in target
Verify data transformation rules are applied correctly
Perform data profiling to identify anomalies
Use checksums or hash values to ensure data integrity
Validation of BI reports involves verifying data accuracy, completeness, and consistency.
Verify data accuracy by comparing with source systems
Ensure completeness by checking all expected data is present
Check consistency by comparing with historical data or benchmarks
Validate calculations and aggregations
Test report functionality and user experience
posted on 17 Oct 2024
I joined inm because of the company's reputation for innovation and growth in the tech industry.
Attracted to company's reputation for innovation
Excited about opportunities for growth in tech industry
Impressed by company culture and values
I applied via Recruitment Consulltant and was interviewed in Feb 2024. There was 1 interview round.
Filter stage is used in ETL processes to selectively pass or reject data based on specified criteria.
Filter stage helps in removing unwanted data from the input dataset.
It can be used to apply conditions like filtering out duplicate records, selecting specific columns, or excluding certain values.
For example, a filter stage can be used to only pass records where the sales amount is greater than $1000.
Transform stage is used in ETL process to apply business rules, clean and enrich data before loading into target database.
Transform stage is used to apply business rules to the data.
It is used to clean and standardize data before loading into the target database.
Transform stage can also be used to enrich data by combining multiple sources or adding calculated fields.
Examples include converting data types, removing dupl
Sort stage is used in ETL processes to sort data based on specified criteria before loading it into the target system.
Sort stage helps in arranging data in a specific order for better analysis and reporting
It can be used to remove duplicates from data before loading
Sorting can be done based on multiple columns or expressions
Example: Sorting customer data based on their purchase amount before loading into a data warehou
To create a parallel job, use parallel processing techniques to divide tasks into smaller subtasks that can be executed simultaneously.
Identify tasks that can be executed independently and in parallel
Use parallel processing techniques such as multi-threading or distributed computing
Implement parallel job using ETL tools like Informatica or Talend
Monitor and optimize parallel job performance to ensure efficient executio
SCD stands for Slowly Changing Dimension, a technique used in data warehousing to track changes in dimension attributes over time.
SCD is used to maintain historical data in a data warehouse
There are different types of SCDs - Type 1, Type 2, and Type 3
Type 1 SCD overwrites old data with new data
Type 2 SCD creates a new record for each change and maintains history
Type 3 SCD keeps a limited history by adding columns to tr
Group data by column 'A', calculate mean of column 'B' and sum values in column 'C' for each group.
Use groupby() function in pandas to group data by column 'A'
Apply mean() function on column 'B' and sum() function on column 'C' for each group
Example: df.groupby('A').agg({'B':'mean', 'C':'sum'})
deepcopy() creates a new object with completely independent copies of nested objects, while copy() creates a shallow copy.
deepcopy() creates a new object and recursively copies all nested objects, while copy() creates a shallow copy of the top-level object only.
Use deepcopy() when you need to create a deep copy of an object with nested structures, to avoid any references to the original object.
Use copy() when you only ...
Python decorators are functions that modify the behavior of other functions. They are commonly used for adding functionality to existing functions without modifying their code.
Decorators are defined using the @ symbol followed by the decorator function name.
They can be used to measure the execution time of a function by wrapping the function with a timer decorator.
Example: def timer(func): def wrapper(*args, **kwargs...
I applied via Naukri.com and was interviewed in Mar 2024. There was 1 interview round.
Cubes are multidimensional structures used in OLAP (Online Analytical Processing) to analyze data from multiple perspectives.
Cubes store aggregated data for faster analysis
They allow users to drill down into data to analyze at different levels of granularity
OLAP cubes are used in business intelligence for complex data analysis
Examples include sales cubes, finance cubes, and customer analysis cubes
ETL pipeline stands for Extract, Transform, Load pipeline used to extract data from various sources, transform it, and load it into a data warehouse.
ETL pipeline involves extracting data from multiple sources such as databases, files, APIs, etc.
The extracted data is then transformed by applying various operations like cleaning, filtering, aggregating, etc.
Finally, the transformed data is loaded into a data warehouse or...
I applied via Approached by Company and was interviewed in Jun 2024. There was 1 interview round.
posted on 9 Sep 2024
Experienced Oracle DBA with 5+ years of hands-on experience in managing databases, optimizing performance, and ensuring data security.
5+ years of experience as an Oracle DBA
Proficient in database management, performance optimization, and data security
Skilled in troubleshooting and resolving database issues
Strong knowledge of Oracle database architecture and SQL
Certified Oracle Database Administrator (OCA/OCP)
The greatest challenge faced during migration to cloud from on-premise DB is ensuring data security and compliance.
Ensuring data security during transit and at rest
Maintaining compliance with regulations and industry standards
Minimizing downtime and ensuring data integrity
Optimizing performance and cost efficiency in the cloud environment
Software Engineer
22.8k
salaries
| ₹1.2 L/yr - ₹8 L/yr |
Technical Lead
20.9k
salaries
| ₹6.9 L/yr - ₹25 L/yr |
Senior Software Engineer
15.6k
salaries
| ₹4 L/yr - ₹16.9 L/yr |
Lead Engineer
14.8k
salaries
| ₹4.2 L/yr - ₹14 L/yr |
Analyst
14k
salaries
| ₹1.2 L/yr - ₹6.7 L/yr |
TCS
Wipro
Accenture
Cognizant