Posted in

Data Platform Engineer-HPC

Data Platform Engineer-HPC

CompanySanofi
LocationToronto, ON, Canada
Salary$93900 – $135633.33
TypeFull-Time
DegreesBachelor’s, Master’s
Experience LevelMid Level, Senior

Requirements

  • Extensive experience managing cloud infrastructure, particularly in AWS (EC2, VPC, FSx for Lustre, EKS, Lambda, RDS, DynamoDB, S3, CloudFront, etc.)
  • Strong understanding of networking concepts, including subnetting, routing, NAT, VPNs, Transit Gateway, BGP, DNS, load balancing, and firewall rule management.
  • Expertise in Infrastructure as Code (IaC) tools, such as Terraform (preferred), CloudFormation, CDK.
  • Experience with containerization and orchestration technologies (Docker, Kubernetes).
  • Experience with working with Schedulers (Slurm, PBS, Grid Engine…) and applications wrappers
  • Strong automation and scripting skills in Python, Bash for infrastructure and operational tasks.
  • Experience with CI/CD pipelines (GitHub Actions, GitLab CI, AWS CodePipeline, Jenkins, ArgoCD, etc.).
  • Knowledge of identity and access management (IAM, SSO, MFA…) for securing cloud environments.
  • Experience in monitoring and logging tools such as CloudWatch, Prometheus, Grafana, Datadog.
  • Strong problem-solving skills with the ability to troubleshoot networking and infrastructure issues across complex environments.
  • Excellent communication skills to collaborate effectively with researchers, DevOps, and IT security teams.
  • Bachelor’s or Master’s degree in Computer Science, Engineering, or a related field.

Responsibilities

  • Design, deploy, and optimize cloud-based computing infrastructures to support scientific research activities.
  • Manage and secure AWS cloud environments (VPC, Transit Gateway, Direct Connect, Route 53) and its integration to the Sanofi environment (Zscaler/firewalls).
  • Implement and troubleshoot network architectures, ensuring high availability, low latency, and security compliance across cloud-based and hybrid infrastructures.
  • Automate infrastructure deployment and management using Infrastructure as Code (IaC) tools such as Terraform, AWS CloudFormation, or CDK.
  • Develop and manage Kubernetes clusters and containerized workflows to facilitate scientific computing workloads.
  • Implement CI/CD pipelines for automating infrastructure updates and application deployments, ensuring minimal downtime.
  • Monitor, optimize, and troubleshoot cloud and hybrid environments, ensuring efficient resource utilization and cost-effectiveness.
  • Manage access control policies ensuring compliance with security best practices.
  • Support hybrid connectivity, and routing configurations across cloud and on-prem environments.
  • Assist researchers in deploying, scaling, and troubleshooting cloud-based applications for computational workloads through HPC schedulers.
  • Stay updated with the latest advancements in cloud networking, security, and DevOps methodologies, recommending improvements to optimize research infrastructure.

Preferred Qualifications

  • AWS Certified Solutions Architect, DevOps Engineer, or Advanced Networking Specialty certifications.
  • Experience with other cloud platforms (Google Cloud, Azure).
  • Familiarity with zero-trust security models and cloud-native security best practices.
  • Experience with scientific computing and R&D workflows (Cellranger, Nextflow, R/RStudio).
  • Hands-on experience with policy-as-code tools (HashiCorp).