Research Engineer, Model Inference & Serving - Paris
Quick Summary
About H:H exists to push the boundaries of superintelligence with agentic AI. By automating complex, multi-step tasks typically performed by humans, AI agents will help unlock full human potential.
Develop scalable, low-latency and cost effective inference pipelines Optimize model performance: memory usage, throughput, and latency, using advanced techniques like distributed computing, model compression, quantization and caching mechanisms…
Technical skills: MS or PhD in Computer Science, Machine Learning or related fields Proficient in at least one of the following programming languages: Python, Rust or C/C++ Experience in GPU programming such as CUDA, Open AI Triton, Metal, etc.
Responsibilities
~1 min readLocation & Eligibility
Listing Details
- Posted
- April 14, 2026
- First seen
- May 6, 2026
- Last seen
- May 13, 2026
Posting Health
- Days active
- 0
- Repost count
- 0
- Trust Level
- 16%
- Scored at
- May 6, 2026
Signal breakdown
Please let hcompany know you found this job on Jobera.
4 other jobs at hcompany
View all →Explore open roles at hcompany.
Similar Research Engineer jobs
View all →Browse Similar Jobs
Stay ahead of the market
Get the latest job openings, salary trends, and hiring insights delivered to your inbox every week.
No spam. Unsubscribe at any time.