Engineer ML
About the role
The mission of the team is to transform terabytes of data into robust models for improving Network Experience and Customer Experience and reducing Operational Expenditure.
Candidate will be required to understand the different Datasets being supplied from different sources and design a production grade big data architecture for Reporting and Analytics Platform. He will be involved in developing in-house algorithms as per Business Requirements. He will also be involved in vendor reviews of solution designs and related code on the principle of adherence to high-quality development principles while delivering solutions on time and on budget.
Key Responsibilities
- Develop best quality software design and Big Data architecture for a Product that is being migrated to Big Data Platform from MYSQL based architecture.
- This product is spread across dimensions of big data, analytics, machine learning, graph problems, etc.
- The person is responsible to design and develop systems that use machine learning, or AI, to build models that can automate processes. They use data to create models, perform statistical analysis, and train systems to improve performance.
- Identify, prioritize and execute tasks in the software development life cycle.
- Develop tools and applications by producing clean, efficient code, automate tasks through appropriate tools and scripting.
- Review and debug code and perform validation and verification testing
- Collaborate with internal teams and external vendors to fix and improve current Architecture that helps to improve software development process and team productivity.
- Document development phases and monitor systems. Guide daily DevOps and ensure timely hardware expansions or architecture revamp.
- Provide statistical analysis to develop, test, and optimize databases to their full potential.
- Ensure software is up-to-date with latest technologies.
- Review Vendor Codes and solution designs for on-time deployment.
Experience & Skills
- Must have experience in Machine Learning and experience in machine learning frameworks like TensorFlow or PyTorch.
- Knowledge of selected programming languages (e.g. Python, JAVA, Scala Python, C++)
- In-depth knowledge of relational databases (e.g. PostgreSQL, MySQL) and NoSQL databases (e.g. MongoDB or Hadoop)
- Hands-on experience working with Big Data technologies like Hadoop, MongodB, Hive, Pig, Oozie, Map Reduce, Spark, Sqoop, Kafka, Flume, etc.
- Extensive experience in software design, development and scripting.
- Experience with real-time streaming tools like Apache Kafka and NiFi is a plus.
- Has hands-on experience working with large volumes of data including different patterns of data ingestion, processing (batch & real-time), movement, storage and access (for both internal and external to BU) and is able to make independent decisions within scope of project. ( Spark, Kafka, Kstream/Flink)
- Very strong analytical skills with the demonstrated ability to research & make decisions based on the day-to-day and complex customer problems
- Understanding of HTTP and RESTful web services is a plus
- Experience with version control tools (SVN or Git).
Educational Qualifications: Bachelor's or Master's Degree in Computer Science or related field; or equivalent related professional experience
Work Experience: 1-7 years of total experience preferably in Big Data
Employment Type: Full Time, Permanent
Read full job description