We are hiring a Staff Data Engineer to join our India Operations and play a crucial role in our mission to establish a world-class data engineering team within the Center for Data and Insights (CDI)
Reporting directly to the Director of Data Engineering, you will be a key contributor, advancing our data engineering capabilities in the AWS and GCP ecosystems
Your responsibilities include collaborating with key stakeholders, guiding and mentoring fellow data engineers, and working hands-on in various domains such as data architecture, data lake infrastructure, data and ML job orchestration
Your contributions will ensure the consistency and reliability of data and insights, aligning with our objective of enabling well-informed decision-making
The ideal candidate will demonstrate an empathetic and service-oriented approach, fostering a thriving data and insights culture while enhancing and safeguarding our data infrastructure
This role presents a unique opportunity to build and strengthen our data engineering platforms at a global level
If you are an experienced professional with a passion for impactful data engineering initiatives and a commitment to driving transformative changes, we encourage you to explore this role
Joining us as a Staff Data Engineer allows you to significantly contribute to the trajectory of our CDI, making a lasting impact on our data-centric aspirations as we aim for new heights
Core Areas of Responsibility
Implement robust data infrastructure, platforms, and solutions.
Collaborate effectively with cross functional teams & CDI leaders, by ensuring the delivery of timely data load and jobs tailored to their unique needs.
Guide & mentor the team of skilled data engineers, by prioritizing a service-oriented approach and quick response times.
Advocate for the enhancement, and adherence to high data quality standards, KPI certification methods, and engineering best practices.
Approach reporting platforms and analytical processes with innovative thinking, considering the evolving demands of the business.
Implement the strategy for migrating from AWS to GCP with near real time events, machine learning pipelines using our customer data platform (Segment) and purpose built pipelines and DBs to activate systems of intelligence.
Continuously improve reporting workflows and efficiency, harnessing the power of automation whenever feasible.
Enhance the performance, reliability, and scalability of storage and compute layers of the data lake.
About You
We get excited about candidates, like you, because...
8+ years of hands-on experience in data engineering and/or software development.
Highly skilled in programming languages like Python, Spark & SQL
Comfortable using BI tools like Tableau, Looker, Preset, and so on
Proficient in utilizing event data collection tools such as Snowplow, Segment, Google Tag Manager, Tealium, mParticle, and more.
Comprehensive expertise across the entire lifecycle of implementing compute and orchestration tools like Databricks, Airflow, Talend, and others.
Skilled in working with streaming OLAP engines like Druid, ClickHouse, and similar technologies.
Experience leveraging AWS services including EMR Spark, Redshift, Kinesis, Lambda, Glue, S3, and Athena, among others. Nice to have exposure to GCP services like BigQuery, Google Storage, Looker, Google Analytics, and so on
Good understanding of building real-time data systems as well AI/ML personalization products
Experience with Customer Data Platforms (CDPs) and Data Management Platforms (DMPs), contributing to holistic data strategies.
Familiarity with high-security environments like HIPAA, PCI, or similar contexts, highlighting a commitment to data privacy and security.
Accomplished in managing large-scale data sets, handling Terabytes of data and billions of records effectively.
You holds a Bachelors degree in Computer Science, Information Systems, or a related field, providing a strong foundational knowledge