02:30 PM - 02:55 PM
From 2005 onwards, Forklift University has been at the forefront of training people in Powered Industrial Trucks (PITs). They do both OSHA-compliant training and issue OSHA certificates.
The brain retains more information from real-life experiences. The millennial generation requires training that captivates and teaches in a more visual manner. So, Forklift University needed a captivating solution for Gen Z while being user-friendly and adaptive for Gen Y to the boomers.
Also, when it comes to PITs training, safety is of utmost importance, followed by cost. They needed a solution that would enable their users to learn the dos and don’ts of forklift driving in a reasonably realistic warehouse setting.
Travancore Analytics (TA) has been a critical player in the extended reality domain. When Forklift University approached TA with the problem, TA suggested a combination of Metaverse and VR to achieve the intended needs. TA created a realistic warehouse and sit-down forklift using 3D software. The same was incorporated into a complex, yet user-friendly application developed using Unity. The application works in conjunction with HTC VIVE Pro 2, a VR head-mounted device. The combination allows the user to drive a virtual forklift in a virtual
environment and undertake training curated with OSHA guidelines in mind. With various training modules of varying degrees of difficulty and functionality, Forklift University can provide its users with close-to real-life forklift training with complete safety. With VR technology, we can create multiple PITs training software under various environments that replicate real-life challenges and incorporate several features that, in turn, develop more in-depth training software.
01:55 PM - 02:20 PM
Construction is a complex and messy process, combining millions of individual components - all needing to be positioned in the exact right place with mm-level accuracy. However, the construction process has largely remained unchanged for thousands of years. A group of people design a building, listening to every wish and elaborate desire coming from the Owner, and then they hand the design over to thousands of other people to actually build it, without any understanding of how this design fits the real world. It’s kind of like building a massive jigsaw puzzle where thousands of people are responsible for one piece or another, and no one really knows how they all fit together. This waterfall process leads to building things up only to tear them down immediately after and then build them right back again - just moved over by 1ft - something the construction industry spends over $280B per year doing. This is simply not sustainable, for the industry, for the stakeholders, and most importantly - for the planet. With nearly 40% of the world’s Carbon emissions being contributed by the construction industry, transformation is desperately needed.
And that’s exactly what Trimble is working to do. As a leader in high-accuracy positioning technologies, Trimble has a long standing history of bringing precision to the construction industry - helping to fit all those puzzle pieces together. But we see the opportunity to do more. Since 1997, when we filed our first XR patent, Trimble has been transforming the way the world works by connecting the physical and digital worlds. Now, we’re working to change this archaic narrative, by empowering everyone in construction to visualize, interpret, and action the design through Augmented and Mixed Reality technologies in the field. From catching design mistakes by viewing 3D models on any worker’s iPad, to being more efficient by controlling the most precise Total Station with nothing more than your gaze, we are improving communication and collaboration around design intent, enabling more efficient and sustainable projects. Follow us on this journey as we outline how Extended Reality technologies are revolutionizing the way the construction industry operates today, tomorrow, and for decades to come.
02:55 PM - 03:20 PM
Doublepoint builds a gesture-based touch interface to enable efficient, robust, and intuitive AR interaction. At this talk, they will showcase how these solutions can work as standalone input modalities or be combined to supercharge eye tracking and hand tracking.
User input is one of the many challenges standing in the way of the mass adoption of AR. How will the everyday person interact in Augmented Reality? What's the role of touch in interaction? At Doublepoint we research different ways for people to interact in AR and develop the best technologies to detect these interactions efficiently and reliably.
Currently, there are a lot of input methods for AR using built-in sensors of the headset, such as hand tracking, eye tracking, and voice input. However, if we want it to be as transformative as the graphical user interface or the capacitive touch screen, we need to put deliberate thought into building the ideal input paradigm and the needed hardware that might not be in the headset itself.
At this talk:
• We’ll demonstrate how a machine learning algorithm on existing smartwatches can already significantly improve AR interaction.
• We’ll show how it can be combined with eye tracking and hand tracking sensors in the headset to improve interactions even more.
• Lastly, we'll show some of our future custom hardware dedicated to sensing advanced micro-gestures in a small and convenient form factor.
11:30 AM - 11:55 AM
Digital Twins connect physical systems with virtual representations and models to allow for visual representations, integration of sensor data, and predictive capabilities for how assets or processes will behave in the future. Globally, organizations are grappling with the acceleration of remote operations and an increasingly prevalent skills gap. Forward-looking companies are addressing this problem by equipping their in-person, hybrid, and off-site teams with mixed reality (MR) solutions that enhance productivity capabilities especially when integrated with Digital Twins. In this session, learn how Sphere and AWS are working together to develop a digital workspace which enables professionals to communicate across languages, distances, dimensions, and time (with predictive capabilities). By partnering on initiatives which include TwinMaker and Lookout for Vision, as well as cloud rendering powered by AWS and its partners, Sphere’s cutting-edge solution is pioneering the future of collaboration, as well as expert assistance and workflow guidance using MR.
09:00 AM - 09:25 AM
Both large and small companies are looking at VR with AR passthrough MR. The use of smaller and smaller optics, particularly systems using pancake optics, seems to have accelerated this trend. Everything that works well with Passthrough MR is a major problem to solve in Optical MR and vice versa. This presentation will discuss the optical, display, and human factor issues associated with Optical MR compared to VR with Passthrough MR.
02:35 PM - 03:30 PM
According to leading executives of XR companies, what are the key considerations when designing haptics experiences and products? What insights and lessons do you need to know to save you from reinventing the wheel? What are the brightest minds in XR predicting the future of haptics will look like?
05:35 PM - 06:00 PM
Behind the world’s first XR creator tool for responsive content, SyncReality is premiering its latest offering at AWE Santa Clara. The Beta release of the SyncReality Suite, will don the main tool, for automated placement of virtual assets in the real world, a room scanner, a parametric asset tool, and a simulator.
The SyncReality Suite is a revolutionary XR spatial content creation tool- enabling any XR content to automatically adapt to the space of the enduser - and designed to unlock new possibilities for businesses, brands, and developers, that want to provide enhanced user experiences across vCommerce, education, gaming, and more - to engage and excite their audiences.
Following the launch of Alpha in February 2023, SyncReality Beta boasts cutting-edge features including:
• Optimized product interface for easy navigation and creation
• Streamlined workflow management for cross-team development
• Improved user experience for seamless interactions
• Parameric Asset Bundles, keeping aesthetics intact in a broad variety of spaces.
• Simulator to verify XR content before exporting.
SyncReality’s unique technology makes it easier than ever for developers to create seamless experiences that engage and excite users. With SyncReality, you can create immersive worlds that allow users to interact with a character, world, brand, or product in ways never before possible. Imagine:
• Enjoying a concert in your living room by your favorite musician or
• Learning a new skill at home while sitting in a classroom environment or
• Playing an interactive escape game that transforms your space into a vast rainforest!
01:30 PM - 01:55 PM
XR cannot scale until...
Luis Ramirez, Mawari’s Founder and CEO, will deliver a comprehensive presentation that covers the delivery bottlenecks and technological advancements in XR development needed for XR delivery to truly scale. He will also showcase case studies from various sectors, including transportation, education, and entertainment, to discuss whether the ubiquitous XR Cloud and ever persistent digital twin of the world are currently achievable. This talk will be organized to provide a clear and concise overview of the topic.