Aller au contenu principal
INRIA recrutement

Post-Doctoral Research Visit F - M Developing 3D-Vlms For Embodied Perception And Reasoning H/F INRIA

  • Paris - 75
  • CDD
  • 12 mois
  • Service public des collectivités territoriales

Les missions du poste

Post-Doctoral Research Visit F/M Developing 3D-VLMs for Embodied Perception and Reasoning
Le descriptif de l'offre ci-dessous est en Anglais
Type de contrat : CDD

Contrat renouvelable : Oui

Niveau de diplôme exigé : Thèse ou équivalent

Fonction : Post-Doctorant

Contexte et atouts du poste

The work will be conducted in the WILLOW team at Inria Paris research center, supported by the French National Research Agency (ANR) JCJC funding. Renowned for its exceptional work in computer vision and robotics, the WILLOW team has consistently produced high-quality research, resulting in publications in major journals and conferences.

As part of the team, you will have access to a well-established laboratory featuring multiple robotic arms, hands, quadrupeds, bipeds, and mobile manipulators.

Additionally, you can expect frequent visits and talks by esteemed researchers from top research laboratories around the world. Opportunities abound for collaboration with leading researchers both in Europe and globally.

Furthermore, you will join an international and welcoming team environment, where we regularly organize various events ranging from casual after-work gatherings to multi-day lab retreats.

Mission confiée

As populations age and labor shortages grow, autonomous robots are increasingly seen as a practical solution in fields ranging from manufacturing and agriculture to healthcare, customer service, and home assistance. In recent years, robotic hardware has made impressive strides. For example, dexterous end-effectors, high-precision sensors, and affordable mobile platforms are now widely available. Yet, the intelligence driving these robots still lags behind. Most deployed systems still operate on rigid, pre-programmed scripts. While this works well in controlled, predictable settings, performance quickly breaks down in unstructured environments where objects, scenes, and tasks vary widely. The inability to generalize across such variations remains one of the most significant bottlenecks in robotics.

To address this challenge, this project aims to develop advanced 3D vision-language models (3D-VLMs) that endow robots with high-level perception and reasoning skills. These models will need to understand complex 3D scenes, perform spatial and commonsense reasoning, long-horizon planning and precise spatial grounding. These capabilities should generalize across different robot embodiments. For example, when instructed to take the shoes out of the shoe box, the model should first detect that the box is closed, reason about how to open it, and predict the location and contact points of the lid. At any stage, a user could intervene, e.g., correcting place the shoe on the left, not the right.

To this end,the candidatewill investigate three main directions: 1) Learning generalizable 3D representations for robotics; 2) Designing new 3D-VLM architectures for reasoning and high-level planning; 3) Improving computational efficiency to make these models practical for real-world deployment.

Collaborators:

Shizhe Chen, Cordelia Schmid, Stephane Caron

Principales activités

- Read papers
- Propose methods
- Conduct experiments
- Analyze results
- Write papers
- Present work in conferences
- Co-supervise students (optional)

Compétences

The candidate must have an excellent track of records and a PhD Degree. The candidate must have the following qualifications:

- Strong background in computer vision, robotics, or related fields

- Excellent programming skills with deep learning using Python and Pytorch

- Strong proficiency in both written and spoken English

- Ability to work independently as well as collaboratively

- Publications in top-tier vision/robotics conferences and contributions to open-source vision/robotic projects are appreciated

Avantages

- Subsidized meals
- Partial reimbursement of public transport costs
- Leave: 7 weeks of annual leave + 10 extra days off due to RTT (statutory reduction in working hours) + possibility of exceptional leave (sick children, moving home, etc.)
- Possibility of teleworking and flexible organization of working hours
- Professional equipment available (videoconferencing, loan of computer equipment, etc.)
- Social, cultural and sports events and activities
- Access to vocational training
- Social security coverage

A propos d'Inria

Inria est l'institut national de recherche dédié aux sciences et technologies du numérique. Il emploie 2600 personnes. Ses 215 équipes-projets agiles, en général communes avec des partenaires académiques, impliquent plus de 3900 scientifiques pour relever les défis du numérique, souvent à l'interface d'autres disciplines. L'institut fait appel à de nombreux talents dans plus d'une quarantaine de métiers différents. 900 personnels d'appui à la recherche et à l'innovation contribuent à faire émerger et grandir des projets scientifiques ou entrepreneuriaux qui impactent le monde. Inria travaille avec de nombreuses entreprises et a accompagné la création de plus de 200 start-up. L'institut s'eorce ainsi de répondre aux enjeux de la transformation numérique de la science, de la société et de l'économie.

Publiée le 13/09/2025 - Réf : 2ea267605dda3ed5d1d51e13ac2634ff

Post-Doctoral Research Visit F - M Developing 3D-Vlms For Embodied Perception And Reasoning H/F

INRIA
  • Paris - 75
  • CDD
Publiée le 13/09/2025 - Réf : 2ea267605dda3ed5d1d51e13ac2634ff

Finalisez votre candidature

sur le site du recruteur

Créez votre compte pour postuler

sur le site du recruteur !

Voir plus d'offres
Les sites
L'emploi
  • Offres d'emploi par métier
  • Offres d'emploi par ville
  • Offres d'emploi par entreprise
  • Offres d'emploi par mots clés
L'entreprise
  • Qui sommes-nous ?
  • On recrute
  • Accès client
Les apps
Application Android (nouvelle fenêtre) Application ios (nouvelle fenêtre)
Nous suivre sur :
Informations légales CGU Politique de confidentialité Gérer les traceurs Accessibilité : non conforme Aide et contact