Set Up Oculus Lipsync for Native Development

The Oculus Lipsync native libraries are used to sync avatar lip movements to speech sounds. Oculus Lipsync analyzes an audio input stream from a microphone input or an audio file, and either offline or in real-time predicts a set of values (called visemes) which may be used to animate the lips of an avatar.

Oculus Lipsync offers a library for native C++ development on Windows or macOS.

The following sections describe the requirements, download and setup for native development with the Oculus Lipsync libraries on Windows or macOS.

Requirements

To use the Lipsync native library you must have a C/C++ compiler installed on your development computer. For example, you could use one of the following IDEs:

  • Microsoft Visual Studio C++ 2015 or later
  • Xcode 8 or later
  • Android NDK 12b or later

Download

Setup

To setup the Lipsync package for Visual Studio or Xcode, choose one of the following:

  • In Visual Studio, access your Project Properties. In the dialog:
    • Under C/C++ > General, for Additional Include Directories, choose Edit and provide the path to the Lipsync\Include directory (under the extraction directory from the previous step). This will be something like [extraction-dir]\LipSync\Native\Include. The following image shows an example:

    • Under Linker > General settings, under Additional Library Directories, choose Edit and specify the path to lib folder that contains OVRLipSyncShim.lib for your target platform to the list of library dependencies. For example, to add the lib file for Windows 64 development, add [extraction-dir]\LipSync\Native\Lib\Win64.

  • In Xcode:
    • Add libOVRLipSyncShim.a to Link Binary with Libraries, and make it Required. The following image shows an example.

    • For Search Paths, add ../OVRLipSync/Native/package/Include folder to Header Search Paths and ../OVRLipSync/Native/package/Lib/MacOs/ folder to Library Search Paths.The following image shows an example.