Build and Maintain data pipelines for PoCs/Analytics/Data Science
Prepare, maintain and optimize data using automation for easier consumption as per teams dynamic requirements
Responsible for data reliability, efficiency, and quality
Manage multiple cloud environments
Research and Develop tools for cloud infrastructure
Collaborate with other teams for data pipeline issues and resolution
Flexible in timings to work with global teams
Provide guidance and mentoring to less- experienced staff members
What you need to bring:
Bachelors or Masters degree in Computer Science (or equivalent)
8- 20 years of experience
Understanding of big data pipelines and distributed computing
Hands-on experience in building or maintaining data solutions using Kafka/Spark/Hadoop/ArangoDB or similar technologies in public and hybrid cloud environments.
Exposure to Visualization tools
Experience working in any of the cloud computing environments such as AWS, Azure and GCP.
Experience in one or more of programming/scripting languages such as Scala/Java/Python.