Job Summary:
Saama Technologies is seeking a skilled GCP Data Engineer with expertise in Python, GCP, SQL, Big Data, and Data Warehousing. The ideal candidate will have 4-10 years of experience in designing, building, and maintaining scalable data pipelines and data warehouses on Google Cloud Platform (GCP). The candidate should be proficient in handling large datasets, optimizing data workflows, and ensuring data quality and integrity.
Key Responsibilities:
- Design and Develop Data Pipelines: Build and maintain scalable, efficient, and reliable data pipelines on GCP using tools like Cloud Dataflow, BigQuery, Cloud Storage, and Pub/Sub.
- Data Warehousing: Design and implement data warehousing solutions using BigQuery and other GCP services to support business intelligence and analytics needs.
- Big Data Processing: Work with large datasets using Big Data technologies such as Apache Beam, Apache Spark, or Hadoop to process and analyze data efficiently.
- SQL Expertise: Write complex SQL queries for data extraction, transformation, and loading (ETL) processes.
- Python Development: Develop and maintain Python scripts for data processing, automation, and integration with GCP services.
- Data Integration: Integrate data from various sources (structured and unstructured) into GCP, ensuring data consistency and quality.
- Performance Optimization: Optimize data pipelines and queries for performance, scalability, and cost-efficiency on GCP.
- Collaboration: Work closely with data scientists, analysts, and other stakeholders to understand data requirements and deliver solutions that meet business needs.
- Data Governance: Implement data governance practices, including data security, privacy, and compliance with industry standards.
- Troubleshooting: Identify and resolve issues related to data pipelines, data quality, and performance.
Required Skills and Qualifications:
- Experience: 4-10 years of experience in data engineering, with a focus on GCP, Python, SQL, and Big Data technologies.
- GCP Expertise: Hands-on experience with Google Cloud Platform (GCP) services such as BigQuery, Cloud Dataflow, Cloud Storage, Pub/Sub, and DataProc.
- Programming: Strong proficiency in Python for data processing and automation.
- SQL: Advanced SQL skills for querying, data transformation, and optimization.
- Big Data: Experience with Big Data tools like Apache Spark, Apache Beam, or Hadoop.
- Data Warehousing: Experience in designing and managing data warehouses, preferably using BigQuery.
- ETL/ELT: Strong understanding of ETL/ELT processes and tools.
- Problem-Solving: Strong analytical and problem-solving skills with the ability to troubleshoot and resolve data-related issues.
- Communication: Excellent communication and collaboration skills to work effectively with cross-functional teams.
- Location: Preferred candidates from Pune, Chennai, and Coimbatore.
Why Join Saama?
- Innovative Environment: Work on cutting-edge technologies and solve complex data challenges.
- Growth Opportunities: Opportunities for professional growth and skill development.
- Collaborative Culture: Be part of a collaborative and inclusive work environment.
- Impact: Contribute to impactful projects that drive business decisions and innovation.
Employment Type: Full Time, Permanent
Read full job description