Member of technical staff (Inference)

Permanent contract
Paris
No remote work
Salary: Not specified

H Company
H Company

Interested in this job?

Questions and answers about the job

The position

Job description

About H: H exists to push the boundaries of superintelligence with agentic AI. By automating complex, multi-step tasks typically performed by humans, AI agents will help unlock full human potential.

H is hiring the world’s best AI talent, seeking those who are dedicated as much to building safely and responsibly as to advancing disruptive agentic capabilities. We promote a mindset of openness, learning and collaboration, where everyone has something to contribute.

Holistic, Humanist, Humble.


About the Team: The Inference team develops and enhances the inference stack for serving H-models that power our agent technology. The team focuses on optimizing hardware utilization to reach high throughput, low latency and cost efficiency in order to deliver a seamless user experience.

Key Responsibilities:

  • Develop scalable, low-latency and cost effective inference pipelines

  • Optimize model performance: memory usage, throughput, and latency, using advanced techniques like distributed computing, model compression, quantization and caching mechanisms

  • Develop specialized GPU kernels for performance-critical tasks like attention mechanisms, matrix multiplications, etc.

  • Collaborate with H research teams on model architectures to enhance efficiency during inference

  • Review state-of-the-art papers to improve memory usage, throughput and latency (Flash attention, Paged Attention, Continuous batching, etc.)

  • Prioritize and implement state-of-the-art inference techniques

Requirements:

  • Technical skills:

    • MS or PhD in Computer Science, Machine Learning or related fields

    • Proficient in at least one of the following programming languages: Python, Rust or C/C++

    • Experience in GPU programming such as CUDA, Open AI Triton, Metal, etc.

    • Experience in model compression and quantization techniques

  • Soft skills

    • Collaborative mindset, thriving in dynamic, multidisciplinary teams

    • Strong communication and presentation skills

    • Eager to explore new challenges

  • Bonuses:

    • Experience with LLM serving frameworks such as vLLM, TensorRT-LLM, SGLang, llama.cpp, etc.

    • Experience with CUDA kernel programming and NCCL

    • Experience in deep learning inference framework (Pytorch/execuTorch, ONNX Runtime, GGML, etc.)

Location:

  • Paris or London.

  • This role is hybrid, and you are expected to be in the office 3 days a week on average.

  • The final decision for this will lie with the hiring manager for each individual role

What We Offer:

  • Join the exciting journey of shaping the future of AI, and be part of the early days of one of the hottest AI startups

  • Collaborate with a fun, dynamic and multicultural team, working alongside world-class AI talent in a highly collaborative environment

  • Enjoy a competitive salary

  • Unlock opportunities for professional growth, continuous learning, and career development

If you want to change the status quo in AI, join us.

These job openings might interest you!

These companies are also recruiting for the position of “Data / Business Intelligence”.

See all job openings