ACCUEILPLATEFORMESTUDIOSECTEURSACADÉMIENOUS
VR GLOSSARY
Definition

LLM (Large Language Model)

AI model trained on vast amounts of text to understand and generate natural language

Lire en français
LLM (Large Language Model)

Explanation

An LLM (Large Language Model) is an AI model capable of understanding and generating natural language text. In XR, LLMs enable intelligent voice assistants integrated into headsets and glasses, generate dialogues for virtual characters, and offer conversational interaction with immersive environments. They transform human-machine interaction by making it natural and contextual.

Real-world example

In a Meta Quest headset, you ask aloud 'Show me how to disassemble this motor' and the AI assistant (powered by an LLM) understands your request, generates the instructions, and displays them in AR overlaid on the actual part.

Practical applications

  • XR voice assistant: asking questions in natural language inside a headset
  • Intelligent NPCs: virtual characters capable of realistic dialogue
  • Content generation: creating training scenarios from simple text descriptions
  • Real-time translation: understanding and translating conversations in AR glasses

LLM use cases in XR

Embedded assistants

  • Conversational AI in glasses
  • Understanding visual context + text
  • Real-time responses

Example: Meta AI integrated into Ray-Ban Meta to answer questions about what you see

Adaptive training

  • Dynamically generated scenarios
  • Personalized natural language feedback
  • Adapts to the learner's level

Example: A virtual trainer that adjusts explanations based on your mistakes

XR content creation

  • Dialogue and script generation
  • Text description to 3D scene
  • Automated localization

Example: Describe a training scenario and let the AI generate the VR experience

VR scenario

A technician wears AR glasses on an industrial site. They look at a faulty valve and ask 'What is the replacement procedure?' The embedded LLM identifies the model via the camera, consults the technical documentation, and displays the steps as an overlay, with voice instructions.

Why it matters in professional VR

  • LLMs make XR interaction natural (voice, text) instead of menus and buttons
  • They enable faster and more cost-effective immersive content creation
  • Embedded AI in glasses is a key differentiator (Meta, Google, Rokid)
  • The combination of LLMs + computer vision unlocks unprecedented AR use cases