This website uses cookies to improve our services and deliver relevant ads.
By interacting with this site, you agree to this use. For more information, see our Cookies Policy

Archived Documentation

This version of the guide is out of date. Click here for the latest version.

Oculus Lipsync Unity Integration Guide

The Oculus Lipsync Unity integration is a tool used to sync avatar lip movements to speech sounds. Oculus Lipsync analyzes an audio input stream from a microphone input or an audio file, and either offline or in real-time predicts a set of values (called visemes) which may be used to animate the lips of an avatar.

This guide describes how to install and use Oculus Lipsync for Unity.