Agenda

  • Expand all
Jun 1

01:00 PM - 01:25 PM

Description

Join Michael Running Wolf and Caroline Running Wolf for a discussion about one of the most wonderful uses cases of XR for good as they presenting findings from their research on how immersive tech and AI can be used to reclaim indigenous languages that would otherwise be lost.

Speakers

Founder & Technical Director , McGill University
PhD Student , University of British Columbia
May 31

04:35 PM - 05:00 PM

Description

With the use of a standard phone and a 3D scan of a location, you can determine your position with an impressive accuracy of 5cm - a game changer for AR. This technology opens doors to a variety of features including AR indoor navigation, AR workflows, and even remote assistance across large facilities. Spatial applications will make AR a tool for daily use, creating massive ROI for businesses. This talk will show how to use AR for daily operations and ultimately how to increase your ROI.

Speakers

CEO , VIEWAR
May 31

05:05 PM - 05:30 PM

Description

Discover the untapped potential of XR technology as we discuss the crucial role of audio in driving realistic 3D experiences. Whether you’re venturing into VR, AR, or MR, delivering an exceptional audio experience is paramount to ensuring active participant engagement and seamless integration within virtual environments.

Join us for an enlightening conversation with industry experts, Paul Boustead, VP of Product at Dolby, and Michael Kolbrener, COO at Odyssey.stream. Gain valuable insights into the Dolby.io offerings that are powering immersive XR landscapes. And learn about the latest advancements and future possibilities that await XR enthusiasts in the near and long term.

Speakers

VP of Product & Architecture , Dolby.io
COO , Odyssey
Jun 2

11:40 AM - 12:05 PM

Description

The advent of advanced technology and machinery has put workers in a variety of industries at risk of unemployment. In order to solve this problem, we will explain a human assist-based XR solution that helps overcome physical limitations (memory, cognition, vision, etc.) as a human being and coexist with technology, focusing on cases. And I will suggest the direction that XR technology should go forward.

Speakers

Associate , MetaVu
Jun 1

05:05 PM - 05:30 PM

Description

How will the evolution to video conferencing look like - and why is bringing people 'together' in virtual spaces not enough? Founder and CEO of Beem, Janosch Amstutz, dives into the technical and psychological factors involved in creating a new communications medium.

Speakers

Founder & Chief Executive , Beem
May 31

02:25 PM - 02:50 PM

Description

The JPEO-CBRND partnered with MRIGlobal and ForgeFX Simulations to produce The CBNRD HoloTrainer, a networked multiuser Microsoft HoloLens 2 augmented reality training simulator for Chemical, Biological, Radiological, and Nuclear detection device operators. This presentation will cover the entire project lifecycle, including government requirements and the development of holographically projected interactive virtual equipment, culminating in deployment to soldiers. The CBRND HoloTrainer is a groundbreaking spatial computing application that significantly increases the effectiveness of its users.

Speakers

Chief Product Officer , ForgeFX Simulations
CEO and Co-Founder , ForgeFX Simulations
Jun 1

02:00 PM - 02:25 PM

Description

First responders (FR) need every advantage during critical moments of fleeting life, and SMART cyber-physical systems (CPS) provide crucial information for time-sensitive events. Wireless real-time convergence of building information modeling, interfacing of 3D data through machine-to-machine and machine-to-human analytics exchange, and bio-telemetry monitoring of victims and responders will save lives.

Auto-populated, legacy, and real-time remote sensed point clouds would allow FR personnel and their commanders to intelligently and visually understand their environments during response. Mixed physics computational fluid dynamic programs allow for predictive outputs in structure and wildland fires, hazardous materials releases, and other natural and human-initiated events. Technical rescues and high-life hazards will benefit from SMART Cities, IoT, remote sensing, virtual, augmented, and mixed realities. The unconscious patient will be treated via facial recognition software, allowing responders to know medical history, and complimentary real-time field communications will be transmitted to base hospital trauma surgeons-in 3D.

The SMART FR- during the response, mitigation, and post-incident analysis, will view emergencies through 3D projected light and holographic displays and immersive HUD lenses. The multi-sensor fusion-enabled technology is focused on saving civilian lives and empowering responders.

Speakers

President , McKinzie Smart Technologies
Fire Chief , San Bernardino County Fire Department
May 31

03:25 PM - 04:20 PM

Description

Coming Soon!

Speakers

Founder | XR Industry Pioneer , Wizard Wells
Administrator - XR Systems , Blue Origin
Technical Fellow , The Boeing Company
Solutions Architect , Amazon Web Services
Jun 1

02:30 PM - 02:55 PM

Description

From 2005 onwards, Forklift University has been at the forefront of training people in Powered Industrial Trucks (PITs). They do both OSHA-compliant training and issue OSHA certificates.

The brain retains more information from real-life experiences. The millennial generation requires training that captivates and teaches in a more visual manner. So, Forklift University needed a captivating solution for Gen Z while being user-friendly and adaptive for Gen Y to the boomers.

Also, when it comes to PITs training, safety is of utmost importance, followed by cost. They needed a solution that would enable their users to learn the dos and don’ts of forklift driving in a reasonably realistic warehouse setting.

Travancore Analytics (TA) has been a critical player in the extended reality domain. When Forklift University approached TA with the problem, TA suggested a combination of Metaverse and VR to achieve the intended needs. TA created a realistic warehouse and sit-down forklift using 3D software. The same was incorporated into a complex, yet user-friendly application developed using Unity. The application works in conjunction with HTC VIVE Pro 2, a VR head-mounted device. The combination allows the user to drive a virtual forklift in a virtual
environment and undertake training curated with OSHA guidelines in mind. With various training modules of varying degrees of difficulty and functionality, Forklift University can provide its users with close-to real-life forklift training with complete safety. With VR technology, we can create multiple PITs training software under various environments that replicate real-life challenges and incorporate several features that, in turn, develop more in-depth training software.

Speakers

President , Forklift University
VP - Engineering , Travancore Analytics
May 31

01:55 PM - 02:20 PM

Description

Construction is a complex and messy process, combining millions of individual components - all needing to be positioned in the exact right place with mm-level accuracy. However, the construction process has largely remained unchanged for thousands of years. A group of people design a building, listening to every wish and elaborate desire coming from the Owner, and then they hand the design over to thousands of other people to actually build it, without any understanding of how this design fits the real world. It’s kind of like building a massive jigsaw puzzle where thousands of people are responsible for one piece or another, and no one really knows how they all fit together. This waterfall process leads to building things up only to tear them down immediately after and then build them right back again - just moved over by 1ft - something the construction industry spends over $280B per year doing. This is simply not sustainable, for the industry, for the stakeholders, and most importantly - for the planet. With nearly 40% of the world’s Carbon emissions being contributed by the construction industry, transformation is desperately needed.
And that’s exactly what Trimble is working to do. As a leader in high-accuracy positioning technologies, Trimble has a long standing history of bringing precision to the construction industry - helping to fit all those puzzle pieces together. But we see the opportunity to do more. Since 1997, when we filed our first XR patent, Trimble has been transforming the way the world works by connecting the physical and digital worlds. Now, we’re working to change this archaic narrative, by empowering everyone in construction to visualize, interpret, and action the design through Augmented and Mixed Reality technologies in the field. From catching design mistakes by viewing 3D models on any worker’s iPad, to being more efficient by controlling the most precise Total Station with nothing more than your gaze, we are improving communication and collaboration around design intent, enabling more efficient and sustainable projects. Follow us on this journey as we outline how Extended Reality technologies are revolutionizing the way the construction industry operates today, tomorrow, and for decades to come.

Speakers

VDC Manager , Canadian Turner Construction Company
Product Manager , Trimble
May 31

02:55 PM - 03:20 PM

Description

Doublepoint builds a gesture-based touch interface to enable efficient, robust, and intuitive AR interaction. At this talk, they will showcase how these solutions can work as standalone input modalities or be combined to supercharge eye tracking and hand tracking.

User input is one of the many challenges standing in the way of the mass adoption of AR. How will the everyday person interact in Augmented Reality? What's the role of touch in interaction? At Doublepoint we research different ways for people to interact in AR and develop the best technologies to detect these interactions efficiently and reliably.

Currently, there are a lot of input methods for AR using built-in sensors of the headset, such as hand tracking, eye tracking, and voice input. However, if we want it to be as transformative as the graphical user interface or the capacitive touch screen, we need to put deliberate thought into building the ideal input paradigm and the needed hardware that might not be in the headset itself.

At this talk:

• We’ll demonstrate how a machine learning algorithm on existing smartwatches can already significantly improve AR interaction.
• We’ll show how it can be combined with eye tracking and hand tracking sensors in the headset to improve interactions even more.
• Lastly, we'll show some of our future custom hardware dedicated to sensing advanced micro-gestures in a small and convenient form factor.

Speakers

Co-Founder and CTO , Doublepoint
May 31

11:30 AM - 11:55 AM

Description

Digital Twins connect physical systems with virtual representations and models to allow for visual representations, integration of sensor data, and predictive capabilities for how assets or processes will behave in the future. Globally, organizations are grappling with the acceleration of remote operations and an increasingly prevalent skills gap. Forward-looking companies are addressing this problem by equipping their in-person, hybrid, and off-site teams with mixed reality (MR) solutions that enhance productivity capabilities especially when integrated with Digital Twins. In this session, learn how Sphere and AWS are working together to develop a digital workspace which enables professionals to communicate across languages, distances, dimensions, and time (with predictive capabilities). By partnering on initiatives which include TwinMaker and Lookout for Vision, as well as cloud rendering powered by AWS and its partners, Sphere’s cutting-edge solution is pioneering the future of collaboration, as well as expert assistance and workflow guidance using MR.

Speakers

Global Business Dev. & GTM Leader , AWS
CEO , Sphere (by holo|one)
Jun 2

09:00 AM - 09:25 AM

Description

Both large and small companies are looking at VR with AR passthrough MR. The use of smaller and smaller optics, particularly systems using pancake optics, seems to have accelerated this trend. Everything that works well with Passthrough MR is a major problem to solve in Optical MR and vice versa. This presentation will discuss the optical, display, and human factor issues associated with Optical MR compared to VR with Passthrough MR.

Speakers

President , KGonTech
Jun 2

02:35 PM - 03:30 PM

Description

According to leading executives of XR companies, what are the key considerations when designing haptics experiences and products? What insights and lessons do you need to know to save you from reinventing the wheel? What are the brightest minds in XR predicting the future of haptics will look like?

Speakers

Business Development Manager , bHaptics
CEO / Co-Founder , Sensoryx
Strategic Partnerships , TITAN Haptics
Vice President, Sales , HaptX
Strategic Partnership Manager , Contact CI
May 31

05:35 PM - 06:00 PM

Description

Behind the world’s first XR creator tool for responsive content, SyncReality is premiering its latest offering at AWE Santa Clara. The Beta release of the SyncReality Suite, will don the main tool, for automated placement of virtual assets in the real world, a room scanner, a parametric asset tool, and a simulator.

The SyncReality Suite is a revolutionary XR spatial content creation tool- enabling any XR content to automatically adapt to the space of the enduser - and designed to unlock new possibilities for businesses, brands, and developers, that want to provide enhanced user experiences across vCommerce, education, gaming, and more - to engage and excite their audiences.

Following the launch of Alpha in February 2023, SyncReality Beta boasts cutting-edge features including:

• Optimized product interface for easy navigation and creation
• Streamlined workflow management for cross-team development
• Improved user experience for seamless interactions
• Parameric Asset Bundles, keeping aesthetics intact in a broad variety of spaces.
• Simulator to verify XR content before exporting.

SyncReality’s unique technology makes it easier than ever for developers to create seamless experiences that engage and excite users. With SyncReality, you can create immersive worlds that allow users to interact with a character, world, brand, or product in ways never before possible. Imagine:

• Enjoying a concert in your living room by your favorite musician or
• Learning a new skill at home while sitting in a classroom environment or
• Playing an interactive escape game that transforms your space into a vast rainforest!

Speakers

Founder / CEO , SyncReality
Jun 1

01:30 PM - 01:55 PM

Description

XR cannot scale until...

Luis Ramirez, Mawari’s Founder and CEO, will deliver a comprehensive presentation that covers the delivery bottlenecks and technological advancements in XR development needed for XR delivery to truly scale. He will also showcase case studies from various sectors, including transportation, education, and entertainment, to discuss whether the ubiquitous XR Cloud and ever persistent digital twin of the world are currently achievable. This talk will be organized to provide a clear and concise overview of the topic.

Speakers

Founder and CEO , Mawari