ARENA User Presence
Users in the ARENA are represented to each other in a 3d scene with an avatar. Your default user avatar looks like a static 3d floating head model with your chosen display name above the head, and a muted microphone icon. Wherever you choose to move or look within a scene is where other users will see your avatar move or look.
Upon entering a scene you will see a panel for audio and video setup. Here you can
Microphone to use, test your devices to make sure they are working, and adjust your display name.
User Display Name
The display name you choose, will be displayed in random colored text above your user avatar. In an ARENA Scene, you can change your display text by clicking the last right-hand arrow “V” icon to show the
Settings panel, which provides a
Display Name input box you can change to what you wish.
If you click the camera icon, you will begin to send video to other users. The video you are sending to other users will appear in the upper left corner of the screen. If your scene is enabled to use the facial recognition feature, you can click the recognition avatar icon and you will see the facial feature points recognized in the upper left corner of your screen.
|User Avatar||Image||Camera||Video Sent|
|Static 3d head (default)||TBD||Off||None|
|Facial recognition 3d head||TBD||On||No video, only facial feature points|
|Video cube||TBD||On||Camera video|
|Black cube||TBD||On||Camera video error, only black frames|
Camera input is not available on all mobile device/browser combinations.
If you click the microphone icon, you will begin to send audio to other users. When you have your microphone turned off, other users will see a muted microphone icon over your avatar. Our audio support includes a number of features. Depending on the range between you and other users’ avatars, you may experience louder sounds the closer you are to them, depending which side of you they are speaking from, and if they are far enough away from you to cut off the audio completely.
|User Audio||User Range||Microphone||Audio Sent|
|Muted Head Icon||all||Off||None|
|Directional Audio||0-20m||On||Audio is in stereo by avatar pose|
|Spatial Audio||0-20m||On||Audio fades with distance|
|Cut Off Audio||> 20m||On||None|
By default the ARENA shows your location as a 3d model of a head, with your nose at your position coordinates. The pose of a user is composed of the user’s position (location) and rotation (gaze) in the 3d world of the ARENA. Depending on the XR view and hardware you use, your pose may be determined by a number of differing sensors.
|XR View||Rotation Sensor||Position Sensor|
|VR Headset/Glasses (VR button)||Accelerometer||———|
|AR Phone/Tablet (AR button)||Accelerometer||Accelerometer/Camera|
Facial Recognition Avatar
Your ARENA scene may have the capability to use facial recognition to recognize your facial feature points from your camera and present a generic 3d head which will move its eyes, eyebrows, mouth and lips as you do. Only other users will see the 3d rendering of your avatar.
When you click and accept the screenshare icon and popup, it will ask for the object name/ID of the object you want to screen share on (defaulted to an object with
screenshare, which is dynamically created if it doesn’t exist already).
- See our Screenshare Walkthrough for more detail.
Vive (laser) controls
Out of Date
We have some support for VR Vive hand controls in the ARENA, but these instructions may be out of date.
We’ve noticed the controllers don’t show up in the scene unless they both and EVERYTHING else for SteamVR are all working (headset, lighthouses). And sometimes you have to restart SteamVR for hand controllers to show up in the scene; even though SteamVR shows them as being working/on/available/etc. It’s possible to open VR mode in an ARENA scene and be missing the hand controls.
By default we use A-Frame
laser-controls which default to showing Valve Index controller 3D models (gray, circular), even if we are using (equivalent) Vive controllers (black, paddle shaped, not included in the list of controllers known to A-Frame).
(from A-Frame documentation)
|The camera component defines from which perspective the user views the scene. The camera is commonly paired with control components that allow input devices to move and rotate the camera.|
|A camera should usually be positioned at the average height of human eye level (1.6 meters). When used with controls that receive rotation or position (e.g. from a VR device) this position will be overridden.|
<a-entity camera look-controls position="0 1.6 0"></a-entity>
The above example puts the camera at a position in the scene, but sure enough, when we use a tablet with WebXR Viewer or a VR or AR headset, these values are overwritten. In fact, it turns out that from a desktop browser, at the start of our A-Frame session, regardless of the values set in the HTML above, the start position is set to (0, 1.6, 0). It was misleading that the HTML definition just happened to match. Our code sets it to (0,0,0) in the declaration. It gets more interesting: on a tablet or phone, the start position again gets overridden by (0,0,0) this time!
|When moving or rotating the camera relative to the scene, use a camera rig. By doing so, the camera’s height offset can be updated by room-scale devices, while still allowing the tracked area to be moved independently around the scene.|
<a-entity id="rig" position="25 10 0"> <a-entity id="camera" camera look-controls></a-entity> </a-entity>