Posted in

Senior Machine Learning Engineer AI Inferencing

Senior Machine Learning Engineer AI Inferencing

CompanyServiceNow
LocationSanta Clara, CA, USA
Salary$158500 – $269500
TypeFull-Time
Degrees
Experience LevelSenior

Requirements

  • Experience in leveraging or critically thinking about how to integrate AI into work processes, decision-making, or problem-solving.
  • Low Latency Optimization: Experience in optimizing models for low latency inference, important for real-time applications.
  • High Throughput Optimization: Knowledge of maximizing inference throughput.
  • Real-time Systems: Understanding the constraints of real-time systems on model inference.
  • Model Quantization and Compression: Practical experience in reducing model size and computational cost.
  • Proficient in prompt engineering and developing LLM based features
  • Experience in using AI productivity tools such as Cursor, Windsurf, etc
  • Minimum 5 years of experience working in Software Development role.
  • Proficiency in Python and Golang, with a strong grasp of software engineering principles.
  • Hands-on experience with prompt engineering: ability to craft, test, and optimize prompts for task accuracy and efficiency.
  • Demonstrated ability to thrive in fast-paced, dynamic environments.
  • Knowledge of unit testing, profiling, and code tuning.

Responsibilities

  • Play a major part in building AI and Machine Learning (ML) platform that transform the user experience and workflow efficiency of enterprise services.
  • Utilize expertise in Python and Golang to develop high-performance components of the AI Platform.
  • Collaborate with cross-functional teams to integrate AI capabilities seamlessly into workflows and user experiences.
  • Ensure reliability and performance of AI models by applying best practices in software engineering and AI inferencing.
  • Stay ahead of the curve by quickly learning emerging technologies and applying them to enhance the AI Platform.

Preferred Qualifications

    No preferred qualifications provided.