Bachelors and/or master s degree in computer science or equivalent experience
Must have total 6+ yrsof IT experience and 3+ years experience in Data warehouse/ETL projects
Deep understanding of Star and Snowflake dimensional modelling
Strong knowledge of Data Management principles Good understanding of Databricks Data & AI platform and Databricks Delta Lake Architecture
Should have hands-on experience in SQL, Python and Spark (PySpark) Candidate must have experience in AWS/ Azure stack Desirable to have ETL with batch and streaming (Kinesis)
Experience in building ETL / data warehouse transformation processes Experience with Apache Kafka for use with streaming data / event-based data Experience with other Open-Source big data products Hadoop (incl
Hive, Pig, Impala) Experience with Open Source non-relational / NoSQL data repositories (incl
MongoDB, Cassandra, Neo4J) Experience working with structured and unstructured data including imaging & geospatial data
Experience working in a Dev/Ops environment with tools such as Terraform, CircleCI, GIT
Proficiency in RDBMS, complex SQL, PL/SQL, Unix Shell Scripting, performance tuning and troubleshoot Databricks Certified Data Engineer Associate/Professional Certification (Desirable)
Comfortable working in a dynamic, fast-paced, innovative environment with several ongoing concurrent projects Should have experience working in Agile methodology Strong verbal and written communication skills
Strong analytical and problem-solving skills with a high attention to detail. Mandatory Skills: Python/ PySpark / Spark with Azure/ AWS Databricks