Staff Data Engineer
Company | ShyftLabs |
---|---|
Location | Toronto, ON, Canada |
Salary | $Not Provided – $Not Provided |
Type | Full-Time |
Degrees | Bachelor’s, Master’s |
Experience Level | Senior |
Requirements
- Bachelor’s or Master’s degree in Computer Science, Engineering, or related technical field
- 6+ years of experience in enterprise data engineering with Fortune 500 companies
- Expert-level experience with Databricks platform including job scheduling and cluster management
- Hands-on experience with Active Batch or similar enterprise job orchestration platforms
- Deep knowledge of AWS data services including DMS, S3, and database connectivity
- Proven experience with CDC (Change Data Capture) implementations at enterprise scale
- Strong background in Delta Live Tables (DLT) and streaming data processing
- Understanding of Apache Spark optimization and performance tuning
- Experience with enterprise database integration and cross-platform data connectivity
Responsibilities
- Implement and optimize enterprise-scale data pipeline architecture processing 1500+ daily jobs across multiple business domains
- Build and maintain Databricks job orchestration solutions integrated with Active Batch scheduling platforms
- Implement data ingestion frameworks supporting CDC via AWS DMS and S3-based data lake patterns
- Troubleshoot and resolve pipeline reliability issues including overrun problems, dependency failures, and compute optimization
- Build observability and monitoring solutions for large-scale data operations teams
- Implement near real-time ingestion patterns using DLT (Delta Live Tables) and streaming architectures
- Build and maintain data quality frameworks integrated with data governance processes
- Implement serverless migration strategies and optimize cloud resource utilization
- Lead technical workshops and provide hands-on guidance to client engineering teams
Preferred Qualifications
- Experience with enterprise integration patterns and API data ingestion
- Knowledge of cloud storage optimization and data lake best practices
- Background in high-volume transaction processing systems
- Experience with retail or energy industry data patterns
- Familiarity with trading and pricing systems data workflows