Understanding in depth both the business and technical problems Dataworks aims to solve
Building tools, platforms and pipelines to enable teams to clearly and cleanly analyze data, build models and drive decisions
Scaling up from laptop-scale to cluster scale problems, in terms of both infrastructure and problem structure and technique
Collaborating across teams to drive the generation of data driven operational insights that translate to high value optimized solutions.
Delivering tangible value very rapidly, collaborating with diverse teams of varying backgrounds and disciplines
Codifying best practices for future reuse in the form of accessible, reusable patterns, templates, and code bases
Interacting with senior technologists from the broader enterprise and outside of FedEx (partner ecosystems and customers) to create synergies and ensure smooth deployments to downstream operational systems
Skill/Knowledge Considered a Plus
Technical background in computer science, software engineering, database systems, distributed systems
Fluency with distributed and cloud environments and a deep understanding of optimizing computational considerations with theoretical properties
Experience in building robust cloud-based data engineering and curation solutions to create data products useful for numerous applications
Detailed knowledge of the Microsoft Azure tooling for large-scale data engineering efforts and deployments is highly preferred. Experience with any combination of the following azure tools: Azure Databricks, Azure Data Factory, Azure SQL D, Azure Synapse Analytics
Developing and operationalizing capabilities and solutions including under near real-time high-volume streaming conditions.
Hands-on development skills with the ability to work at the code level and help debug hard to resolve issues.
A compelling track record of designing and deploying large scale technical solutions, which deliver tangible, ongoing value
Direct experience having built and deployed robust, complex production systems that implement modern, data processing methods at scale
Ability to context-switch, to provide support to dispersed teams which may need an expert hacker to unblock an especially challenging technical obstacle, and to work through problems as they are still being defined
Demonstrated ability to deliver technical projects with a team, often working under tight time constraints to deliver value
An engineering mindset, willing to make rapid, pragmatic decisions to improve performance, accelerate progress or magnify impact
Comfort with working with distributed teams on code-based deliverables, using version control systems and code reviews
Ability to conduct data analysis, investigation, and lineage studies to document and enhance data quality and access
Use of agile and devops practices for project and software management including continuous integration and continuous delivery
Demonstrated expertise working with some of the following common languages and tools:
Spark (Scala and PySpark), Kafka and other high-volume data tools
SQL and NoSQL storage tools, such as MySQL, Postgres, MongoDB/CosmosDB
Java, Python data tools
Azure DevOps experience to track work, develop using git-integrated version control patterns, and build and utilize CI/CD pipelines
Working knowledge and experience implementing data architecture patterns to support varying business needs
Experience with different data types (json, xml, parquet, avro, unstructured) for both batch and streaming ingestions
Use of Azure Kubernetes Services, Eventhubs, or other related technologies to implement streaming ingestions
Experience developing and implementing alerting and monitoring frameworks
Working knowledge of Infrastructure as Code (IaC) through Terraform to create and deploy resources
Implementation experience across different data stores, messaging systems, and data processing engines
Data integration through APIs and/or REST service
PowerPlatform (PowerBI, PowerApp, PowerAutomate) development experience a plus