UE5.3: Roomscale and stationary boundaries break my Player Start! Why?
Hello, I'm having some strange problems. While I try out my test level with the Meta Quest at my workplace and in the Unreal Engine, everything works fine. Now I have created an APK file and installed it on the quest and have the following problem (see also the picture): 1. in the editor "PlayerStart" is at the position and rotation as shown in the picture (see number 1 in the picture). This also works when I test it in the engine and with Meta Quest. 2. However, if I start the level natively on the quest (at a different place in my room), then I start somewhere else in the level and not at the set "PlayerStart" (see number 2 in the picture). And my triggers etc. no longer work. If I reset the position ingame (via menu button), then the player jumps back to the spawnpoint set in the editor (see number 3 in the picture), but the triggers etc. do not work either. If I start the "game" on my workplace, then the spawn position is correct and everything else also works. The spawn point should always be the same, no matter where I start the game from. I don't know why this won't work. New Update: I have found that when I have a stationary boundary (3x3 feet), it works as it should. But as soon as I set Roomscale (or when I am standing), I have the behavior as described above. I'm trying to create a game like "Moss", for example. I have a stationary VR character (first person) where I can interact with my hands and at the same time I control a 3rd person character. And with triggers etc. I then "teleport" the VR character to a different position. This works, but not as soon as I stand or activate Roomscale. And if I'm not in the center of the stationary boundary, then the rotation and position where I "teleport" the VR character is no longer correct. Does anyone have any idea?2.4KViews1like4CommentsHow to get the desk position data from the Room setup process ?
Hello, I have been using Unity for a few months, I am currently in the process of creating a tech demo for the META quest pro, I am using hand tracking and controllers for the room setup process. Currently the desk is just spawned into the scene within the guardian, boundary data. I have managed to call the room setup process but could not find anything in the Oculus documentation and pulling the desk position data from the room/desk setup process. Could anyone provide any tips or suggestions as to how I could get the desk position data and ensure my desk game object spawns at that location? Thanks !!2.3KViews0likes1CommentRoom scale matching IRL location - How to manage Quest2 orientation?
Hi all, I'm creating an app for the Meta Quest2, where a user can walk in and out a room ( L shaped with an open door, nothing fancy ), which has been built based on a real life location, so the virtual environment ( built in Maya using technical drawings of the place ) match the IRL environment. Now I'm having the following issue and I would like some advices: While doing the room scale, I tracked both the inside of the room and a little bit outside the room ( in front of the door ) and set the pawn start location outside the room, at a specific distance from the door, so IRL I also have the same distance as the starting point of the pawn....as soon as the app starts, the pawn has a predefined orientation ( always looks towards the door ), while if I start the app and I'm not looking towards the door IRL, I have a misalignement between the virtual and IRL locations. To temporarily solve this issue I created a 3d printed stand for the Quest2, so that it sit thight there, and I placed the stand in front of the door at a specific distance ( the stand is also placed on top of a pedestal to be able to have the Quest2 at eye level ), so that it matches the distance between IRL and virtual start location. In order to sort out the orientation, I used one of the buttons on the controller to turn on the visibility of the room scale tracking area ( show/hide Guardian Visibility ), so that while the Quest2 is still on the stand, I can look inside the Quest2 and physically rotate the stand to match the orientation between the Guardian tracking area and the virtual environment, using the Guardian tracking area as a guide. When the two environment seems to be aligned, I get the Quest2 from the stand and test it by walking inside the room, and with still the Guardian visibility turned on, I check if the two envirnoments aligned properly. It works, but is not very user friendly, and if the stand is not locked in place after the orientation alignment, the entire calibration is messed up. So is there a proper/different way to align/orient the Quest2 so that the IRL environment match with a virtual environment which has the same geometry? Thanks in advance.8KViews0likes12CommentsHow to view the tracker view cone(s)
I've written up a small post on how to render the Oculus Tracker(s) view cone so you can see the current play space. http://talesfromtherift.com/tracker-view-cone-revisited/ It also shows you how to display a simple camera 3d model too. All credit to Oculus for adding the TrackerAnchor to OVRCameraRig which makes this whole process so much easier! Peter577Views1like0Comments