Posted in

Senior Data Engineer

Senior Data Engineer

CompanyWave Financial
LocationToronto, ON, Canada
Salary$Not Provided – $Not Provided
TypeFull-Time
Degrees
Experience LevelSenior

Requirements

  • 6+ years of experience in building data pipelines and managing a secure, modern data stack
  • At least 3 years of experience working with AWS cloud infrastructure, including Kafka (MSK), Spark / AWS Glue, and infrastructure as code (IaC) using Terraform
  • Write and review high-quality, maintainable code that enhances the reliability and scalability of our data platform, using Python, SQL, and dbt extensively
  • Prior experience building data lakes on S3 using Apache Iceberg with Parquet, Avro, JSON, and CSV file formats
  • Experience with Airflow or similar orchestration systems to build and manage multi-stage workflows that automate and orchestrate data processing pipelines
  • Familiarity with data governance practices, including data quality, lineage, and privacy, as well as experience using cataloging tools to enhance discoverability and compliance
  • Experience developing and deploying data pipeline solutions using CI/CD best practices to ensure reliability and scalability
  • Working knowledge of tools such as Stitch and Segment CDP for integrating diverse data sources into a cohesive ecosystem

Responsibilities

  • Designing, building, and deploying the components of a modern data stack, including CDC ingestion (using Meltano or similar tools), a centralized Iceberg data lake, and a variety of batch, incremental, and stream-based pipelines
  • Helping build and manage a fault-tolerant data platform that scales economically while balancing innovation with operational stability by maintaining legacy Python ELT scripts and accelerating the transition to dbt models in Redshift, Snowflake, or DataBricks
  • Collaborating within a cross-functional team in planning and rolling out data infrastructure and processing pipelines that serve workloads across analytics, machine learning, and GenAI services
  • Independently identifying opportunities to optimize pipelines and improve data workflows under tight deadlines
  • Responding to alerts and proactively implementing monitoring solutions to minimize future incidents, ensuring high availability and reliability of data systems
  • Providing technical assistance and communicating effectively with stakeholders to address their concerns
  • Assessing existing systems, optimizing data accessibility, and providing innovative solutions to help internal teams surface actionable insights that enhance external customer satisfaction

Preferred Qualifications

  • Knowledge and practical experience with Athena, Redshift, or Sagemaker Feature Store to support analytical and machine learning workflows is a definite bonus!