Rapid Motion-Capture-Powered VR Development
Read about the initial release of the A-Frame Motion Capture Components and A-Saturday-Night. Or head directly to the A-Frame Motion Capture Components.
On the A-Frame team, we focus our development on experimentation and innovation with the HTC Vive. But room scale VR can be cumbersome to develop. Every change to the code, we have to:
- Open a web page (often running on a separate computer)
- Enter VR
- Put on the headset
- Grab the controllers (often having to turn them back on)
- Do our test run with the headset and controllers
- Take off the headset and controllers and pop open the browser development tools
- Restart the browser if necessary since they’re currently experimental and buggy
Room scale VR development becomes molasses. But we’ve come up with a workflow to supercharge VR development so we can automate, develop rapidly, and on the go: A-Frame Motion Capture Components.
With the motion capture components, we can record VR actions (e.g., headset and controller movement, controller button presses), and repeatedly replay those VR actions, on any device from anywhere without a headset.
Table of Contents
Motion Capture Makes VR Developer Lives Easier
How to Record
How to Replay
Try It Out!
Below are several real-life use cases of motion capture vastly improving VR development ergonomics:
Faster test trials: No need to take the headset on and off, enter VR, grab the controllers, do manual actions, or restart browsers. Just record once and develop for hours on a single recording.
Development on the go: Rather than having to re-enter the headset and VR every time we want to test something, we can take our recording, send it to, say, a Macbook, head out to a coffee shop, and continue developing our VR application using the recording on a stable browser. Add some
console.logs, refactor your application, or freeze the replay with the A-Frame Inspector (
<ctrl> + <alt> + i) to poke around.
Automated testing: We can record a bunch of different recordings as regression test cases and QA. Store the recordings, do some development, and occasionally click through the recordings to make sure everything still works. We store multiple recordings in projects for later testing. For example, with the Teleport Controls component, I recorded a couple to later test:
Multiple developers sharing one headset: One developer can take a recording with the Vive and go off somewhere else to develop with the recording, leaving the Vive free for the other developers to use or take recordings.
Requests for recordings: Perhaps you don’t have a Vive or Rift handy but your colleague or friend does. Give them a link to your web application maybe via ngrok (isn’t the Web awesome?), have them take a recording, and send it to you! Now you’re unblocked from developing.
Demonstrating bugs: Or let’s say you found a bug in a VR web application and want to show it to someone. Take a recording and send it to them to debug. No need to give bug reproduction steps, it’s all in the recording!
Another use case for motion capture is as an animation tool to create paths and splines, but we won’t go over that here.
Read the Motion Capture documentation for more information. Here’s how to set up the recording:
- Drop the Motion Capture Components script tag into your HTML file (e.g.,
- Add the
avatar-recordercomponent to the scene (i.e.,
- Enter VR
<space>to start recording
- Record movements and actions
<space>to stop recording
- Save the recording JSON file or upload it by hitting
uto get a short URL to share between computers
Now we can replay the recording. You can try recording the camera with WASD and
mouse drag controls right now on a desktop. Head to the Record
Example, open the
browser Console to get feedback, and hit
<space> to start and stop recording!
By default, the recording will also be saved and replayed from localStorage. If we want to take our recording on the go, here’s how to replay a recording (assuming we already have the script tag above):
- Put the recording file somewhere accessible to the web page (i.e., in the project directory or online)
avatar-replayercomponent to the scene (i.e.,
?avatar-recording=path/to/recording.jsonto the URL or set
<a-scene avatar-replayer="src: path/to/recording.json>
Then replay the recording on any device from anywhere without a headset to your heart’s content. Get in the headset, record some clicks, and then from a laptop, we can build event handlers on top of the controller events we emitted in the recording.
The A-Frame Motion Capture Components have a Spectator Mode feature that can be
enabled from the
This lets you view the recording from a third-person view. I found this immensely useful because it was hard to see what was happening in first-person. The first-person view is naturally shaky, hands would occlude the camera, actions would be happening off-screen, and its hard to focus on one place if the camera is always moving away. The Spectator Mode lets you free move around the scene and view at whatever angle or focus on whatever area.
Below is a video of me on Twitch. I was developing room scale VR with A-Frame for almost 3 hours off of one recording. That was some good mileage as I made a ton of progress. I even served my project using ngrok.io so that people could visit the live state of my project with the recording! See motion capture powered development in action: