Posted in

Senior Data Engineer

Senior Data Engineer

CompanyGoosehead Insurance
LocationLakewood, CO, USA
Salary$Not Provided – $Not Provided
TypeFull-Time
Degrees
Experience LevelSenior

Requirements

  • Experience in SQL optimization and performance tuning, and development experience in programming languages like Python, PySpark, Scala etc.
  • Experience working with cloud data platforms (e.g., AWS Redshift, Snowflake, BigQuery).
  • Strong knowledge of data modeling, data warehousing, and building ETL/ELT pipelines.
  • Experience with version control systems like Github and deployment & CI tools.
  • Familiarity with modern data orchestration tools (e.g., Airflow, dbt).
  • Excellent communication and collaboration skills.

Responsibilities

  • Design, build, and maintain scalable ETL/ELT pipelines to support analytics and machine learning workloads
  • Manage and scale data pipelines from internal and external data sources
  • Develop and manage robust data models and warehouse structures that support self-service analytics and reporting
  • Build and own the automation and monitoring frameworks that captures metrics and operational KPIs for data pipeline quality and performance
  • Implement and monitor data quality and validation checks to maintain trust in our data assets
  • Work with stakeholders across the business to understand data requirements and ensure data availability, accuracy, and usability
  • Optimize data storage and query performance across cloud-based and relational systems.
  • Stay current with emerging data engineering tools and architectures
  • Responsible for implementing best practices around systems integration, data modeling, optimization, security, performance, and data management to ensure reliability, scalability, and maintainability of the infrastructure
  • Empower the business by creating value through the increased adoption of data, data science, and business intelligence landscape
  • Collaborate closely with data science and analytics teams to ensure data infrastructure supports model training and deployment
  • Research in state-of-the-art methodologies
  • Create documentation for learning and knowledge transfer
  • Create and audit reusable packages or libraries
  • Implement data security and compliance measures

Preferred Qualifications

  • Experience with real-time data streaming technologies (e.g., Kafka, Kinesis).
  • Familiarity with CI/CD for data pipelines and infrastructure-as-code (e.g., Terraform).
  • Experience supporting machine learning workflows and model deployment.
  • Background in insurance, financial services, or other highly regulated industries.
  • Experienced in coaching and mentoring team members to foster growth and development.