These patents describe systems that use shared visual attention, location, or biometrics to create ad-hoc communication channels and shared digital experiences.
Patents: 1, 4, 6, 9, 10
The patent describes a system for augmenting user experiences—such as viewing a scene or listening to audio—by providing real-time digital annotations. It identifies specific elements like text, objects, faces, or musical themes and retrieves associated information from a database. A unique feature of the system is that it also identifies and lists other users who are simultaneously interacting with the same or similar content.
The inventor identifies that viewing new places, objects, or people in daily life often creates unanswered queries. Similarly, listening to unfamiliar languages or melodies initiates questions regarding the perceived information or music. Until now, there has been no universal system to provide real-time additional information that could change our perception of the world and improve how we learn, communicate, and work.
The system uses a Tracking Unit to capture the central part of a scene or a specific audio segment, which a Conversion Program then identifies with a unique keyword using techniques like OCR, facial recognition, or biometric voice identification. A Management Program receives this keyword and accesses a Database to retrieve both associated annotations (text, images, or video) and a list of other concurrent users. These annotations are then delivered via an Annotations Outlet, such as a smartphone screen or an optical head-mounted display, where they move in coordination with the subject being viewed.
1. Public Social Tagging
Users can "tag" people, landmarks, or objects with public annotations—such as contact info, trivia, or media links—visible to anyone viewing the same subject.
2. Professional Collaboration
In industrial or medical settings, specialists can anchor vocal or visual instructions to specific components, synchronizing these notes across the team's headsets in real-time.
3. Private Personal Notes
Users can store "personal-use" annotations accessible only to themselves, such as private study notes pinned to a physical textbook or price comparisons pinned to products in a store.
4. Privacy-Controlled Sharing
When creating content, users can restrict visibility to specific IDs, allowing private group commentary among friends or colleagues that remains hidden from the general public.
5. Conversational Creation
Users can add or refine annotations using natural language, such as gazing at an object and speaking a command to attach a digital "wish list" tag or a technical memo.

The invention is a system that allows viewers of digital content to engage in real-time social interaction on their primary display. By utilizing AI to analyze scenes and generate descriptive keywords, the system synchronizes viewer feedback with the media timeline. This creates an integrated social layer that enables users to share reactions and comments across different devices and timeframes seamlessly.
Traditional social viewing requires "second-screen" interaction, where users must look away from their TV or headset to type on a phone or tablet. This requirement for a separate device is cumbersome and disrupts immersion, especially for users engaged in "real-world" activities like walking or those using immersive VR headsets. Furthermore, standard social media comments are often disorganized and lack exact synchronization with the specific scene a user is watching.
The system connects directly to the primary display and uses a multi-layered AI approach to bridge the social gap: it first analyzes each scene to generate descriptive keywords tied to the content’s context, then identifies individual viewers through a camera or microphone to associate their unique IDs with their feedback. This feedback is stored in a database and later retrieved as an interactive on-screen overlay for others watching similar content, enabling cross-temporal interaction so viewers can see reactions from people who experienced the same scene in the past.
1. "TV & Primary Display Chatroom"
Viewers can join live, scene-specific discussion groups directly on their TV screen, chatting with others watching the same program without picking up a phone.
2. Immersive VR Social Viewing
Users in virtual reality can see and hear social feedback from friends integrated into their 360-degree environment, preventing the need to remove the headset to check messages.
3. Real-Time Fan Polling & Voting
Reality TV or news programs can capture viewer reactions or votes through the primary display’s microphone, showing live results as an on-screen graphic.
4. Asynchronous "Watch Parties"
A user watching a movie on-demand can see the reactions and "time-stamped" notes left by their friends who watched it days or weeks earlier.
5. Interactive Advertising Feedback
Advertisers can receive instant, scene-specific feedback from viewers via the display’s interface, allowing for real-time optimization of ad campaigns.

This patent describes a system for initiating context-aware communication sessions by combining a user's geographic location with a specific sensor-detected identifier. Using multi-modal recognition (Object, Face, Voice, Scene, etc.), the system validates the convergence of a physical zone and a specific entity to instantiate shared digital environments like chat rooms or collaborative workspaces. These sessions are spatially anchored to the physical world, transforming objects and people into persistent communication hubs.
Prior AR systems focused primarily on retrieving static data, such as digital notes or annotations, associated with a keyword. There remained a critical need for a more dynamic, session-based architecture that supports real-time, synchronous interaction. Furthermore, existing systems lacked strict validation protocols for secure, location-specific collaboration, which is essential for modern applications in retail, enterprise, and social networking.
The system transforms physical entities into "Communication Anchors" by requiring a dual-validation trigger: the user's geographic zone and a specific physical identifier. Once both are verified, the system instantiates a secure digital environment—such as a chatroom or video channel—accessible only to verified users in that same zone. This architecture supports "Contextual Filtering" for different session types (e.g., social vs. professional) and "Historical Persistence," allowing users to view messages left by previous visitors to that same location.
1. "Shop the Look" Interactive Media
Users gaze at items in videos to instantly receive purchase links and metadata.
2. Industrial Maintenance and Repair
Technicians receive real-time AR schematics and safety warnings anchored to machine parts.
3. Hands-Free Medical Diagnostics
Surgeons view 3D anatomical overlays and vitals by shifting their gaze during procedures.
4. Smart Tourism and Education
Users receive historical context for monuments or instant definitions for complex terms in text.
5. Automated Productivity Analysis
Professionals look between files or charts to trigger instant side-by-side data comparisons.

The patent discloses a "Gaze-Bridged Social Mesh" that establishes ad-hoc communication channels between users based on shared visual attention. By tracking gaze vectors via AR/VR headsets or mobile devices, the system detects "dwell events" on specific or semantically linked objects. A Correlation Engine then creates "Visual Lobbies"—temporary audio or data channels—to bridge disparate software environments or physical locations based on what users are looking at.
Current social VR and AR platforms rely on spatial proximity (standing near someone) or manual groupings like party invites to facilitate communication. This creates fragmented user bases isolated within specific applications or servers. For example, a user viewing a virtual car in a racing game currently has no way to spontaneously communicate with another user viewing the same model in a separate showroom application.
The system establishes a semantic bridge between users by calculating the correlation between their gaze targets across different platforms. It utilizes a "Semantic Alignment Processor" to match metadata from different taxonomies, ensuring that two users looking at "similar" objects are grouped together. To ensure intentionality and privacy, the system requires a multi-modal "handshake" (such as gaze combined with a gesture) before opening a channel and offers an "Incognito Gaze" mode to prevent unwanted discovery.
1. Cross-Platform Metaverse Socializing
Enables users in different virtual worlds to chat if they are looking at identical or related digital assets.
2. Interactive E-Commerce Lobbies
Shoppers looking at the same product in different online stores are placed in a shared "Visual Lobby" to discuss the item.
3. Augmented Reality Spectator Sports
Fans in a stadium can be instantly connected to a shared data channel simply by fixating their gaze on the same athlete.
4. Asynchronous "Ghost Rooms"
Allows a user to see and interact with the "gaze-trail" or comments left by previous visitors who looked at the same physical object.
5. Enterprise Collaborative Design
Engineers across different geographic sites can synchronize their focus on specific components of a 3D model, triggering instant collaborative sessions.

The patent discloses a system for creating a "Sympathetic Social Mesh" by connecting users through real-time physiological synchrony. It aggregates biometric data (e.g., heart rate variability, galvanic skin response) from various wearable devices and normalizes them into standardized "Physiological State Vectors" (PSVs). A pattern-matching algorithm identifies groups exhibiting simultaneous biological alignment to trigger context-aware social or mechanical actions.
Current wearable technologies primarily focus on the "Quantified Self," where biometric data is siloed for individual health or fitness tracking. This data remains underutilized for social synthesis, failing to leverage objective internal states to identify connections between people. Consequently, social networking still relies on explicit user inputs or geographic proximity rather than verifiable, "visceral" compatibility.
The system creates a cross-device infrastructure that time-aligns and compares physiological data streams from heterogeneous wearables. By establishing individual user baselines, it detects "Physiological Synchrony"—shared biological responses to an environment or event—without compromising privacy. This synchrony acts as a digital handshake, triggering application-layer commands such as facilitating social discovery or unlocking cooperative game mechanics based on shared internal states.
1. "Visceral" Social Discovery
Connecting individuals in public or virtual spaces based on objective, real-time biological alignment rather than profile algorithms.
2. Cooperative Game Mechanics: Unlocking specific in-game abilities or content only when players reach a state of synchronized physiological focus or excitement.
3. Enterprise Workflow Management: Utilizing group synchrony or stress patterns to optimize notification timing and workload distribution across a team.
4. Crowd Safety and Sentiment Analysis: Detecting anomalies in physiological patterns across a crowd to identify potential safety risks or measure collective engagement at events.
5. Biometric "Ice-Breakers"
Enhancing professional networking or dating apps by highlighting "Physiological State Vector" matches between users in the same vicinity.

Copyright © 2025 HoloVu - All Rights Reserved.