Quest Runtime Optimizer cannot analyze frame capture on Quest Pro v79.1029
I'm getting errors analyzing captures, and after a day spent debugging, it seems to have started with the latest headset update. An untouched project that previously had no problems has the error now after the headset update. There seems to be an error caused by format incompatibility with the trace_processor_shell version (v46.0) and the Core SDK 70+ which causes the python script to crash? (Don't ask me the details, I'm just a tech artist trying to get frames benchmarks here). I'm running the following: Unity 6.1 and 6.2 on Windows 11 Quest Pro (updated to v79.1029 on Aug 25, the date the error started appearing) Here is the error text: MetricAPI: 68AE8CCB.json execode: 0 single positional indexer is out-of-bounds Traceback (most recent call last): File "C:\open\fbsource\arvr\projects\integrations\Unity\RuntimeOptimizer\Editor\PerformanceInsight\MetricAPI.py", line 1118, in main File "pandas\core\indexing.py", line 1191, in __getitem__ File "pandas\core\indexing.py", line 1752, in _getitem_axis File "pandas\core\indexing.py", line 1685, in _validate_integer IndexError: single positional indexer is out-of-bounds UnityEngine.Debug:LogError (object) Meta.XR.RuntimeOptimizer.Core.Util:DebugLogError (object) (at ./Library/PackageCache/com.meta.xr.runtimeoptimizer@e44a64bf5ef6/Core/LogUtils.cs:19) Meta.XR.RuntimeOptimizer.Editor.PerformanceInsight.MetricAPI:GetCaptureMetric (string,string) (at ./Library/PackageCache/com.meta.xr.runtimeoptimizer@e44a64bf5ef6/Editor/PerformanceInsight/MetricAPI.cs:68) Meta.XR.RuntimeOptimizer.Editor.PerformanceInsight.CaptureTool:ProcessCaptureToJsonObjFunc (object) (at ./Library/PackageCache/com.meta.xr.runtimeoptimizer@e44a64bf5ef6/Editor/PerformanceInsight/CaptureTool.cs:787) System.Threading._ThreadPoolWaitCallback:PerformWaitCallback ()30Views0likes0CommentsSuggestion for Developing an SDK for Meta Ray-ban Glasses
Hey everyone, I'm curious: will there ever be a developer kit (SDK) available for Meta glasses? I think the glasses are great, but there's not much to do with them right now. So, I thought, why not build a community around it? If Meta releases an API or something that other apps could call or interact with indirectly (since it seems there's already a service running in the background), this could enable commands from third-party apps. This way, you could allow users to develop their own apps and further customize the experience. For example, let's say I want to do something customizable, like turning on my smart lights using the command "Hey Meta". My own app could register a command (or more than one) so that whenever I say "Hey Meta, [do something]", this command triggers (it could be a post or a deeplink) that my app could receive and then perform the desired action, only for that command. And if this API allows me to use the camera live, like we do in Instagram livestreams, the possibilities would be endless. Think about it, everyone. I really want to build custom experiences with this, so please consider releasing an SDK for Meta glasses. Thanks!61KViews41likes29CommentsS2S REST request not working on some users
So I have a backend for my game and I want to call some S2S requests to get user's data, purchaes and verify purchases. I am using the end point https://graph.oculus.com/${APP_ID}/viewer_purchases to get user's purchases. I provide access_token, user_id, and fields. Just like it mentions in the documention for IAP here https://developers.meta.com/horizon/documentation/native/ps-iap-s2s. The requests are working just fine with some users, but sometimes i enter a user id and get this error: here is an example of it working: if i just enter a random user id: (translation: it says thats an invalid user id) This is not just for this user, it happens for many users. Also it is happing for other requests like https://graph.oculus.com/${USER_ID} to get verified org_id and in consuming and verifying requests. I know that some of these requests dont work for meta test users (users that meta itself uses to test the game) but these users are real players. If anyone has any idea why this is happening please let me know. Side note: the consume_entitlement request also doesn't work sometimes where it throws a 500 is_transient error with code 2 but it still consumes the purchase correctly. weird...44Views0likes0CommentsMQDH Won’t Connect Quest 3 at Final Cable Step Despite USB and ADB Working
Hello everyone, I’m encountering a frustrating issue with Meta Quest Developer Hub (MQDH) v5.7.2 on macOS while trying to set up a Quest 3 running firmware V78. ISSUE DESCRIPTION I enabled Developer Mode on Quest 3 via the mobile app. In the Quest headset, I selected “Always allow USB access.” SideQuest and ADB (adb devices) can successfully detect the Quest 3. However, when I run “Set Up New Device” in MQDH, it stalls at the final step asking me to plug in the cable, even after doing exactly that. No prompt appears inside the headset; MQDH hangs without progressing. I tried reinstalling MQDH and rebooting both Mac and headset—problem persists. WHAT I’VE TRIED Verified Developer Mode is enabled, and USB Connection Dialog permission is granted. Confirmed USB‑C cable works (SideQuest and ADB both connect). Rebooted Quest and Mac multiple times. Uninstalled and reinstalled MQDH, cleared caches/settings. Used both bundled Quest cable and a known good USB‑C data cable (I tested USB 3.2 cable / Thunderbolt cable with Mac mini USB‑C Thunderbolt port). adb works fine. REQUESTED HELP Could someone please advise on potential causes why MQDH stalls even though ADB and SideQuest work? Has anyone else experienced this on Quest 3 running firmware V78 with MQDH 5.7.2? Are there any logs or advanced diagnostics I can capture to pinpoint the blockage? If anyone has successfully connected MQDH after the V78 update, what steps resolved it? FULL ENVIRONMENT INFO Headset Firmware: Quest 3 V78 MQDH Version: 5.7.2 (macOS) OS: macOS (latest version) USB cable: Verified data-capable USB‑C ADB Output: adb devices correctly lists the headset (adb cmd / MacDroid) MQDH Behavior: stuck at final prompt, no progress NEXT STEPS I’m considering doing a factory reset and strictly following the community-suggested flow (skip Air Link until first USB prompt), but not sure if firmware V78 changed behavior. Thank you in advance!80Views0likes0CommentsTechnical Questions for a LBS Game: Disabling System Gestures, Spatial Mapping & Remote Control API
Hello, I'm looking to create a multi-user, large-scale, location-based (offline) game and have a few questions: 1. Is there a way to disable system-level gestures to prevent players from accidentally exiting the application and returning to the home screen during gameplay? 2. Is there a method for scanning a large physical space (approximately 10x10 meters) to generate a persistent and shareable map file? 3. Is it possible to enable or provide some control API? We need an interface that allows a central controller to remotely start and stop the application on all devices, as well as manage the installation and updating of game content.51Views0likes1CommentExploring ChatGPT Integration for VR UI/UX: Has Anyone Tried It?
Hi everyone, I'm currently experimenting with ways to integrate conversational AI (like ChatGPT) into VR environments to enhance user onboarding, contextual help, or dynamic NPC interactions. I’m working with Unity and targeting Quest 2/3. My goal isn’t to run the model locally, but to connect to external APIs (e.g., OpenAI) via secure HTTPS requests — ideally triggered by user input or voice in immersive menus. I’d love to hear from other devs: Have you tested anything similar in a VR context? What are your thoughts on latency and user flow when connecting to external services in VR? Any advice on best practices for handling text responses (display, formatting, user focus)? Are there any sample projects or SDKs that play well with Meta Quest platforms for this use case? This is still a prototype phase for me, but I believe natural language interaction could really improve the way users explore VR content — especially for onboarding or educational apps. Looking forward to your insights!132Views0likes3CommentsPing data overlay
Hi. Fairly new to development and I am stuck with a project. I am trying to develop a program that displays Network information to the user (ping, jitter, packet loss etc.) . I would like this to be able to run over any 3d application in real time Similar to how OVRMetric displays headset information. Any documentation or assistance would greatly be helpful.828Views0likes1CommentAdvanced GPU OVR metrics not working
I've been using OVR Metrics HUD with the advanced GPU metrics since it was made available, and for the most part it has worked very well. Lately though it just stopped working and is giving me garbage values, and I've tried factory resetting, downloading older versions of it, enabling different settings in different orders and nothing is working for me! It's something that I really need to be able to work efficiently so if anyone could help out it'd be amazing!2.4KViews1like4CommentsAdd Chinese input method for Chinese players
Can you add Chinese input method for Chinese players? Now more and more Chinese players have purchased quest3 and quest3s. Although a series of troublesome operations such as VPN are required, the quest equipment is very good, and more and more players are joining the quest. However, more and more Chinese players have reported that the system does not have a Chinese input method, which has a great impact on social aspects, such as Bilibili, the largest video platform in China. I found that YouTube VR has a Chinese input method, so this should not be difficult for developers. I hope developers can add a Chinese input method for Chinese players465Views0likes0CommentsVirtual 3D world anchored to real-world landmarks
## Introduction In an era where immersive technologies have struggled to gain widespread adoption, we believe there is a compelling opportunity to rethink how users engage with digital content and applications. By anchoring a virtual world to the physical environment and seamlessly integrating 2D and 3D experiences, we could create a platform that offers enhanced productivity, intuitive interactions, and a thriving ecosystem of content and experiences. We build upon our previous vision for an AR virtual world by introducing an additional key capability - virtual identity augmentation. This feature allows users to curate and project their digital personas within the shared virtual environment, unlocking new dimensions of social interaction, self-expression, and the blending of physical and virtual realms. ## Key Concepts The core of our proposal revolves around an AR virtual world that is tightly integrated with the physical world, yet maintains its own distinct digital landscape. This environment would be anchored to specific real-world landmarks, such as the Pyramids of Giza, using a combination of GPS, AR frameworks, beacons, and ultra-wideband (UWB) technologies to ensure consistent and precise spatial mapping. Within this virtual world, users would be able to interact with a variety of 2D and 3D elements, including application icons, virtual objects, and portals to immersive experiences. As we previously described, the key differentiator lies in how these interactions are handled for 2D versus 3D devices: 1. **2D Interactions**: When a user with a 2D device (e.g., smartphone, tablet) interacts with a virtual application icon or object, it would trigger an animated "genie out of a bottle" effect, summoning a 2D window or screen that is locked to a fixed position in the user's view. 2. **3D Interactions**: For users with 3D devices (e.g., AR glasses, VR headsets), interacting with a virtual application icon or object would also trigger the "genie out of a bottle" effect, but instead of a 2D window, it would summon a 3D portal or window that the user can physically move around and even enter. ## Virtual Identity Augmentation One of the key new features we are proposing for the AR virtual world is the ability for users to place virtual objects, like hats, accessories, or digital avatars, on themselves. These virtual objects would be anchored to the user's position and movements, creating the illusion of the item being physically present. The critical distinction is that 2D users (e.g., on smartphones, tablets) would be able to see the virtual objects worn by other users in the shared virtual world, but they would not be able to place virtual objects on themselves. This capability would be reserved for 3D device users, who can leverage the spatial awareness and interaction capabilities required for virtual object placement. These virtual objects placed on a user would persist across devices and sessions, creating a consistent virtual identity or "avatar" for that user within the AR virtual world. This virtual identity would be visible to all other users, regardless of their device capabilities (2D or 3D). Importantly, the virtual objects used to create this virtual identity could also be leveraged to partially or completely obscure a user's real-world appearance from 2D video, photo, and 3D scanning. This would allow users to control how they are represented and perceived in the blended physical-virtual environment, providing greater privacy and security. ## Enhanced 2D Interfaces for 3D Users Building on our previous concept, we can further enhance the user experience for 2D applications, particularly for 3D users. By leveraging the depth and spatial characteristics of the 3D interface blocks, we can unlock new ways for users to interact with and manage their virtual applications and content. Some of the key capabilities include: 1. **Contextual Controls and Information Panels**: The sides of the 3D interface blocks could display shortcut controls, supplementary information panels, and other contextual elements that 3D users can access and interact with as they navigate around the application window. 2. **Dynamic Layouts and Customization**: 3D users would be able to resize, rotate, and reposition the side panels and controls, enabling personalized layouts and ergonomic arrangements tailored to their preferences and workflows. 3. **Multi-Dimensional Interactions**: The 3D interface blocks could support advanced interaction methods beyond basic clicking and scrolling, such as gestures (grabbing, pinching, swiping) and voice commands to interact with the contextual controls and information. 4. **Seamless Transition between 2D and 3D**: Despite these enhanced capabilities for 3D users, the 2D application windows would still function as regular 2D interfaces for users without 3D devices, maintaining a seamless collaborative experience across different device types. ## Potential Benefits and Use Cases The enhanced AR virtual world concept we propose offers several potential benefits and use cases: 1. **Increased Productivity and Ergonomics**: By providing 3D users with enhanced controls, contextual information, and customizable layouts, we can improve their efficiency and ergonomics when working with 2D applications. 2. **Intuitive Spatial Interactions**: The ability to physically move and interact with 3D portals and windows, as well as the option to place virtual objects on oneself, can lead to more natural and immersive ways of engaging with digital content and applications. 3. **Virtual Identity and Self-Expression**: The virtual identity augmentation system allows users to curate and project their digital personas, enabling new forms of social interaction, status signaling, and even monetization opportunities. 4. **Privacy and Security**: The option to obscure one's real-world appearance through virtual identity augmentation can provide users with greater control over their digital privacy, especially in public spaces. 5. **Collaborative Experiences**: The seamless integration of 2D and 3D interactions within the same virtual environment can enable users with different device capabilities to collaborate on tasks and projects. 6. **Extensibility and Customization**: Providing tools and APIs for developers to integrate their own applications and content into the virtual world can foster a thriving ecosystem of experiences. 7. **Anchored to the Real World**: Tying the virtual world to specific real-world landmarks can create a sense of spatial awareness and grounding, making the experience feel more meaningful and connected to the user's physical environment. Robotics Safety Integration Real-time visualization of robot operational boundaries Dynamic safety zone mapping visible to all platform users Automated alerts for boundary violations Integration with existing robotics control systems Unified space mapping for multi-robot environments Environmental Monitoring Visualization of invisible environmental factors Air pollution particle mapping CO2 concentration levels Temperature gradients Electromagnetic fields Real-time data integration from environmental sensors Historical data visualization for trend analysis Alert systems for dangerous condition levels Construction and Infrastructure Real-time 3D blueprint visualization Infrastructure mapping Electrical wiring paths Plumbing systems HVAC ducts Network cables Safety feature highlighting for drilling and renovation Progress tracking and documentation Client visualization tools for project understanding Augmented safety checks and compliance monitoring Inventory and Asset Management AI-powered real-time inventory tracking Integration with camera-based stock management systems 3D spatial mapping of warehouse spaces Automated photogrammetry for stock visualization Real-time update of virtual inventory models Cross-reference with ordering systems Predictive analytics for stock management ## Conclusion By combining the core concepts of an AR virtual world with the added capability of virtual identity augmentation, we believe we can create a compelling platform that addresses the shortcomings of past immersive technology efforts. This vision not only offers enhanced productivity, intuitive interactions, and a thriving ecosystem, but also unlocks new dimensions of social interaction, self-expression, and the blending of physical and virtual realms. Creating a shift toward a 3D society, by including 2D phones. Leading to a new 3D app store. We invite you to explore this concept further and consider its potential impact on the future of computing and human-computer interaction. Together, we can shape a new era of spatial computing that bridges the gap between the physical and digital worlds.