In one embodiment, a method includes accessing visual signals comprising images portraying textual content in a real-world environment associated with a first user from a client system associated with the first user, recognizing the textual content based on machine-learning models and the visual signals, determining a context associated with the first user with respect to the real-world environment based on the visual signals, executing tasks determined based on the textual content and the determined context for the first user, and sending instructions for presenting execution results of the tasks to the first user to the client system.
In one embodiment, a method includes receiving, from a client system associated with a user, a first user request that includes a reference to a target object and one or more of an attribute or a relationship of the target object. Visual data including one or more images portraying the target object may then be accessed, and the reference may be resolved to the target object portrayed in the one or more images. Object information of the target object that corresponds to the referenced attribute or relationship of the first user request may be determined based on a visual analysis of the one or more images. Finally, responsive to receiving the first user request, the object information of the target object may be stored in a multimodal dialog state.
The present disclosure provides systems and methods for optimizing media content. One step of the method may include receiving, via a user interface, an indication of a selected post including media content associated with a user. Another step of the method may include evaluating, via a LLM model trained on training data, a quality of the media content of the selected post. A further step may include generating, via the trained LLM model and based upon the evaluated quality, a modified post including modified media content. A change in the modified post is of a first type when the evaluated quality is at or below a threshold or of a second type when the evaluated quality is above a threshold. Even a further step may include transmitting, via the user interface, the modified post for consideration by the user. Yet even a further step may include receiving, via the user interface, an indication of a rejection or an acceptance of the modified post.
In some embodiments, a computer-implemented method includes ascertaining a multitier topology representation of an edge cloud network; generating a pseudo node topology representation of the edge cloud network from the multitier topology representation; and utilizing the pseudo node topology representation of the edge cloud network to ascertain minimum-latency pseudo-node-based edge cloud clusters (ECCs), the minimum-latency pseudo-node-based ECCs being utilized to minimize a latency of user requests routed through the edge cloud network from a user of the edge cloud network. In some embodiments of the computer-implemented method, the minimum-latency pseudo-node-based ECCs are ascertained based upon a pseudo-node-based round-trip-times (RTTs) assessment from the user of the edge cloud network, the user requests being routed to the minimum-latency pseudo-node-based ECCs ascertained using the pseudo node topology representation.
H04L 41/122 - Découverte ou gestion des topologies de réseau des topologies virtualisées, p. ex. les réseaux définis par logiciel [SDN] ou la virtualisation de la fonction réseau [NFV]
Aspects of the present disclosure are directed to an artificial intelligence (“AI”) application running in conjunction with an artificial reality (“XR”) space. The AI Builder responds to user commands, verbal or gestural, to build or edit spaces or objects in space. If the requested object is of a type recognized by the AI Builder, then the AI Builder builds the object from one or more stored templates. The new object's location is determined by the objects that already exist in the user's XR environment and on commands or gestures from the user. If the AI Builder does not recognize the requested object, the user can show an image to the AI Builder, and the AI builds a 3D object in the XR space according to that image. To ease collaboration among users, the AI Builder may present its user interface as a non-player character within the XR world.
G06V 10/70 - Dispositions pour la reconnaissance ou la compréhension d’images ou de vidéos utilisant la reconnaissance de formes ou l’apprentissage automatique
A system and method for providing screen-aware adjustment visualizations associated with displays are provided. The system may facilitate training a large language model to generate a plurality of types of content to improve text content associated with at least one document. The system may further facilitate extracting and reconstructing items of content from a portable document format document. The system may further maintain a semantic order of elements associated with the content. The system may further dynamically organize the content within a display of a device.
A system for establishing trust of entities associated with a system based on social graphs is provided. The system may access a first social graph including at least a first node. The first node is associated with a first set of edges and a first set of neighboring nodes associated with the first set of edges. The system may access a first signature based on the first social graph. The system may receive a request from a second node to establish a trust relationship. The system may access a second signature determined based on a second social graph in response to receiving the request. The system may determine a similarity level between the first signature and the second signature. The system may generate an indication of approval or denial of the request based on the similarity level.
An embodiment includes generating, using a first set of parameters, a first quantization matrix. An embodiment includes encoding, using the first quantization matrix, a frame in an uncompressed video stream, the encoding generating a compressed video stream corresponding to the uncompressed video stream.
H04N 19/172 - Procédés ou dispositions pour le codage, le décodage, la compression ou la décompression de signaux vidéo numériques utilisant le codage adaptatif caractérisés par l’unité de codage, c.-à-d. la partie structurelle ou sémantique du signal vidéo étant l’objet ou le sujet du codage adaptatif l’unité étant une zone de l'image, p. ex. un objet la zone étant une image, une trame ou un champ
H04N 19/196 - Procédés ou dispositions pour le codage, le décodage, la compression ou la décompression de signaux vidéo numériques utilisant le codage adaptatif caractérisés par le procédé d’adaptation, l’outil d’adaptation ou le type d’adaptation utilisés pour le codage adaptatif étant spécialement adaptés au calcul de paramètres de codage, p. ex. en faisant la moyenne de paramètres de codage calculés antérieurement
H04N 19/463 - Inclusion d’information supplémentaire dans le signal vidéo pendant le processus de compression par compression des paramètres d’encodage avant la transmission
9.
METHODS, APPARATUSES AND COMPUTER PROGRAM PRODUCTS FOR GAZE-DRIVEN ADAPTIVE CONTENT GENERATION
Systems and methods are provided for generating adaptive content. The system may implement a machine learning model including training data pre-trained, or trained in real-time based on captured content or prestored content associated with gazes of users, pupil dilations, facial expressions, muscle movements, heart rates, or gaze dwell times of users determined previously or in real-time. The system may determine a gaze(s) of an eye of a user or facial features of a face associated with the user viewing, by a device, items of content in an environment. The system may include determining, based on the gaze(s) or facial features, a state(s) or interest(s) of the user. The system may determine, by implementing the machine learning model and based on the state(s) or interest(s) of the user, content to generate a modification of the items of content or to generate new content items associated with the items of content.
Systems, apparatuses and methods provide technology that receives different numbers of accesses to a data shard from different regions, where the data shard is a portion of a dataset. The technology identifies, with a machine learning model, access patterns based on the different numbers of accesses, and generates, with the machine learning model, values for the different regions based on the access patterns, where the values represent ranks of the different regions that correspond to a future number of predicted accesses to the data shard from the different regions. The technology determines a subset of the different regions to store the data shard based on the values and stores a replica of the data shard in each of the subset of the regions without scarifying latency.
G06F 16/27 - Réplication, distribution ou synchronisation de données entre bases de données ou dans un système de bases de données distribuéesArchitectures de systèmes de bases de données distribuées à cet effet
G06F 11/34 - Enregistrement ou évaluation statistique de l'activité du calculateur, p. ex. des interruptions ou des opérations d'entrée–sortie
11.
Display screen or portion thereof with a transitional graphical user interface
Aspects of the present disclosure are directed to applying voice controls to a target virtual object in an artificial reality environment. User controls in an artificial reality environment can take many forms. Some user controls, such as ray casting or gaze tracking, can incorporate selection mechanics to select the artificial reality environment element (e.g., virtual object) that the user is targeting for interaction. Other forms of user controls, such as voice controls, may not include such selection mechanics. Implementations disambiguate user voice input to select a target virtual object and control the target virtual object based on the voice input. For example, a disambiguation and control layer can select the virtual object the user intends to target with voice input, format input for the target virtual object using the voice input, and control the target virtual object via execution of one or more applications that manage the virtual object.
A system and method for determining content to recommend to a user interface are provided. The system may determine contexts of users within environments. The system may implement a machine learning model including training data pre-trained, or trained in real-time based on historical interactions of users with data, or determined interactions with content by the users in real time. The system may analyze an item(s) of context information associated with the contexts to determine content relevant to a user associated with the system capturing content items within an environment. The system may analyze the item(s) of context information or other items of context information to determine contextual variables, of the environments, determined as relevant to the system. The system may utilize the determined content relevant to the user and the contextual variables determined as relevant to the system to determine a recommendation(s) or action(s) to present to a user interface.
Methods and systems for a fiber optic assembly to propagate light into a waveguide associated with a photonic integrated circuit are provided. The system may include a fiber optic, and a fiber optic core for directing light into the fiber optic assembly. The fiber optic assembly may include at least one transparent layer, metal layer, electrode layer, or a liquid crystal layer. The fiber optic may be aligned in a photonic integrated circuit, where an active feedback loop may be configured to control regions of the fiber optic assembly individually based on the potential difference at both sides of the liquid crystal layer, via electrode layers. The molecules of the liquid crystal layer may be configured to move, change, or be reoriented to direct light into the waveguide based on the potential difference.
G02F 1/13 - Dispositifs ou dispositions pour la commande de l'intensité, de la couleur, de la phase, de la polarisation ou de la direction de la lumière arrivant d'une source lumineuse indépendante, p. ex. commutation, ouverture de porte ou modulationOptique non linéaire pour la commande de l'intensité, de la phase, de la polarisation ou de la couleur basés sur des cristaux liquides, p. ex. cellules d'affichage individuelles à cristaux liquides
15.
Generating Proactive Reminders for Assistant Systems
In one embodiment, a method includes receiving a user request from a first user to create a reminder at a client system, wherein the user request does not specify an activation-condition for the reminder, determining proactive activation-conditions for the reminder, determining whether the proactive activation-conditions for the reminder are satisfied based on user context associated with the first user, and presenting the reminder to the first user at the client system responsive to determining the proactive activation-conditions are satisfied.
In some embodiments, a metaverse optimization and prioritization enabled cloud-based controller includes a metaverse traffic classification unit; and a metaverse optimization and prioritization unit, wherein based upon the identification and classification of data packets as metaverse data packets, the metaverse optimization and prioritization unit optimizes and prioritizes a metaverse client device and metaverse traffic associated with a metaverse optimization and prioritization enabled network. In some embodiments, the metaverse optimization and prioritization unit optimizes the metaverse client device based upon a quality of experience associated with the metaverse client device. In some embodiments, the metaverse optimization and prioritization unit prioritizes the metaverse traffic based on quality of service management features ascertained utilizing the metaverse optimization and prioritization enabled cloud-based controller.
H04W 28/02 - Gestion du trafic, p. ex. régulation de flux ou d'encombrement
H04L 41/0631 - Gestion des fautes, des événements, des alarmes ou des notifications en utilisant l’analyse des causes profondesGestion des fautes, des événements, des alarmes ou des notifications en utilisant l’analyse de la corrélation entre les notifications, les alarmes ou les événements en fonction de critères de décision, p. ex. la hiérarchie ou l’analyse temporelle ou arborescente
H04L 41/5067 - Mesures de la qualité du service [QoS] centrées sur le client
H04W 84/12 - Réseaux locaux sans fil [WLAN Wireless Local Area Network]
17.
Display screen or portion thereof with a graphical user interface
The application describes method of modifying an image. The method may include a step of receiving, via an user interface of a service, a reference image and an input including text associated with the reference image. The method may also include a step of determining, via a trained machine learning (ML) model, one or more features of the reference image. The method may further include a step of modifying, via one or more trained latent diffusion models (LDMs), the reference image based upon the determined features and the received input. Any one or more of a background of the reference image, an area of the reference image or a style of the reference image may be modified. The method may even further include a step of causing to display, via the user interface of the service, the modified image.
G06T 11/60 - Édition de figures et de texteCombinaison de figures ou de texte
G06T 3/40 - Changement d'échelle d’images complètes ou de parties d’image, p. ex. agrandissement ou rétrécissement
G06T 5/50 - Amélioration ou restauration d'image utilisant plusieurs images, p. ex. moyenne ou soustraction
G06V 10/44 - Extraction de caractéristiques locales par analyse des parties du motif, p. ex. par détection d’arêtes, de contours, de boucles, d’angles, de barres ou d’intersectionsAnalyse de connectivité, p. ex. de composantes connectées
19.
CONTEXTUAL IMAGE GENERATION IN MESSAGING APPLICATIONS
A system or method for contextual image generation in messaging applications may include textual input being analyzed using natural language processing to determine an intent to generate an image without requiring an explicit command. An image may be generated based on the determined intent and conversation context, then displayed within the messaging application. The system supports collaborative image refinement, allowing multiple users in a group chat to modify and animate generated images through natural language interactions. A suggestion mechanism may be employed to maintain user control over image generation.
Head-mounted displays may include a machine translation model designed to recognize text through optical character recognition or automatic speech recognition, and may translate the text from its original language to another language. The machine translation model may be trained to modify source text using various tasks, thus allowing the machine translation model to learn different versions of the source text in several different versions. The source text and a variation(s) derived from a task(s) may be mapped to a target text, representing the properly translated and formatted version of the source text. The machine translation model may provide a single model, to facilitate machine translation, implemented on the head-mounted display. Also, the machine translation model may include a bilingual machine translation model that may translate source text from one language to another language, and vice versa.
G06F 40/58 - Utilisation de traduction automatisée, p. ex. pour recherches multilingues, pour fournir aux dispositifs clients une traduction effectuée par le serveur ou pour la traduction en temps réel
Various systems, methods, and devices are described for AI platform that may utilize a machine learning model configured to generate one or more overlays associated with a received input. In an example, systems and methods of generating one or more overlays may include receiving a media item and an input. The input may be natural language text or audio associated with a user. The machine learning model may be used to determine context associated with the input. Based on determining the context of the input, the machine learning model may generate one or more overlays. The user may select from the one or more overlays, indicating one or more overlays to user in conjunction to the media item. A combined media may be provided to the user where the selected overlays may be superimposed on the media item.
Technology herein provides a method, apparatus and computer readable storage medium for use in video encoding. The technology performs operations including pruning interframe candidate modes, based on one or more criteria, to provide a reduced set of candidate modes for encoding a video block, wherein a candidate mode includes an interframe mode type, a set of reference frame types, and one or more dynamic reference list (DRL) candidates, and wherein pruning interframe candidate modes comprises excluding one or more interframe mode types, determining a rate distortion (RD) cost for each of the candidate modes in the reduced set of candidate modes, selecting a candidate mode from the reduced set of candidate modes, based on the lowest RD cost, as a selected interframe mode, and encoding the video block using the selected interframe mode.
H04N 19/109 - Sélection du mode de codage ou du mode de prédiction parmi plusieurs modes de codage prédictif temporel
H04N 19/117 - Filtres, p. ex. pour le pré-traitement ou le post-traitement
H04N 19/139 - Analyse des vecteurs de mouvement, p. ex. leur amplitude, leur direction, leur variance ou leur précision
H04N 19/154 - Qualité visuelle après décodage mesurée ou estimée de façon subjective, p. ex. mesure de la distorsion
H04N 19/176 - Procédés ou dispositions pour le codage, le décodage, la compression ou la décompression de signaux vidéo numériques utilisant le codage adaptatif caractérisés par l’unité de codage, c.-à-d. la partie structurelle ou sémantique du signal vidéo étant l’objet ou le sujet du codage adaptatif l’unité étant une zone de l'image, p. ex. un objet la zone étant un bloc, p. ex. un macrobloc
H04N 19/52 - Traitement de vecteurs de mouvement par encodage par encodage prédictif
23.
Smart Character Suggestion via XR Cubic Keyboard on Head-mounted Devices
In one embodiment, a method includes receiving a first user input from a user from a client system comprising a head-mounted extended-reality (XR) device, determining the user's intent to activate an XR cubic keyboard based on the first user input, rendering the XR cubic keyboard via XR displays of the head-mounted XR device, wherein the XR cubic keyboard comprises input areas representing respective characters in a three-dimensional (3D) space, and wherein the input areas are reachable by respective vectors from a centroid of the XR cubic keyboard in the 3D space, receiving a second user input comprising a hand movement of the user along a direction of a first vector from the centroid of the XR cubic keyboard in the 3D space, determining a first character that the user intended to input, and rendering an indication of the first character via the XR displays.
G06F 3/04886 - Techniques d’interaction fondées sur les interfaces utilisateur graphiques [GUI] utilisant des caractéristiques spécifiques fournies par le périphérique d’entrée, p. ex. des fonctions commandées par la rotation d’une souris à deux capteurs, ou par la nature du périphérique d’entrée, p. ex. des gestes en fonction de la pression exercée enregistrée par une tablette numérique utilisant un écran tactile ou une tablette numérique, p. ex. entrée de commandes par des tracés gestuels par partition en zones à commande indépendante de la surface d’affichage de l’écran tactile ou de la tablette numérique, p. ex. claviers virtuels ou menus
G06F 3/01 - Dispositions d'entrée ou dispositions d'entrée et de sortie combinées pour l'interaction entre l'utilisateur et le calculateur
G06F 3/03 - Dispositions pour convertir sous forme codée la position ou le déplacement d'un élément
G06F 3/0346 - Dispositifs de pointage déplacés ou positionnés par l'utilisateurLeurs accessoires avec détection de l’orientation ou du mouvement libre du dispositif dans un espace en trois dimensions [3D], p. ex. souris 3D, dispositifs de pointage à six degrés de liberté [6-DOF] utilisant des capteurs gyroscopiques, accéléromètres ou d’inclinaison
G06F 3/0482 - Interaction avec des listes d’éléments sélectionnables, p. ex. des menus
G06F 3/0484 - Techniques d’interaction fondées sur les interfaces utilisateur graphiques [GUI] pour la commande de fonctions ou d’opérations spécifiques, p. ex. sélection ou transformation d’un objet, d’une image ou d’un élément de texte affiché, détermination d’une valeur de paramètre ou sélection d’une plage de valeurs
In one embodiment, a method includes receiving a first user input comprising a wake word associated with an assistant xbot from a first client system, setting the assistant xbot into a listening mode, wherein a continuous non-visual feedback is provided via the first client system while the assistant xbot is in the listening mode, receiving a second user input comprising a user utterance from the first client system while the assistant xbot is in the listening mode, determining the second user input has ended based on a completion of the user utterance, and setting the assistant xbot into an inactive mode, wherein the non-visual feedback is discontinued via the first client system while the assistant xbot is in the inactive mode.
In one embodiment, a method includes rendering a first output image of an XR assistant avatar within a first environment associated with a first XR application for displays of a first extended-reality (XR) display device, wherein the XR assistant avatar has a first form according to a first rendering specification associated with the first XR application and the XR assistant avatar is interactable by a first user to access an assistant system, receiving an indication that the first user is switching from the first XR application to a second XR application, accessing a second rendering specification associated with the second XR application, and rendering a second output image of the XR assistant avatar within a second environment associated with the second XR application for displays of a second XR display device, wherein the XR assistant avatar is rendered to have a second form according to the second rendering specification.
Methods, systems, and storage media for running unified simulations on clusters. Exemplary implementations may include: receiving simulation parameters for a simulation of a cluster; generating synthesized workload events based the simulation parameters of the cluster; determining a memory latency associated with the cluster; determining a reliability and availability of resources in the cluster for a predetermined duration of time; simulating events for jobs in the cluster based on the reliability and availability of resources in the cluster, each job associated with one or more synthesized workload events; and outputting simulation results based on the synthesized workload, the memory latency, and the events.
A method of protecting a risk threat is disclosed. A new entry to be stored in a database store is received, wherein the new entry identifies a risk threat. A plurality of disinformation entries is generated based on the new entry to be stored in the database store. Security signatures for the new entry and the plurality of disinformation entries are determined. An authorized user is allowed to use the security signatures to identify the new entry in the database store as a legitimate entry.
G06F 21/57 - Certification ou préservation de plates-formes informatiques fiables, p. ex. démarrages ou arrêts sécurisés, suivis de version, contrôles de logiciel système, mises à jour sécurisées ou évaluation de vulnérabilité
G06F 21/62 - Protection de l’accès à des données via une plate-forme, p. ex. par clés ou règles de contrôle de l’accès
G06F 21/64 - Protection de l’intégrité des données, p. ex. par sommes de contrôle, certificats ou signatures
28.
Reducing Energy Consumption in Extended Reality Devices
A method implemented by a computing device includes displaying on a display of the computing device an extended reality (XR) environment, and determining a context of the XR environment. Determining the context includes determining one or more characteristics associated with a virtual content and a scene of real-world content included within the displayed XR environment. The method further includes determining, based on the one or more characteristics, that the virtual content is displayed prominently with respect to the scene of real-world content, and in response to determining that the virtual content is displayed prominently with respect to the scene of real-world content, forgo an image processing of the scene of real-world content.
Systems and methods for extracting data from digital messages for digital group compositions may include (1) determining that a user of a social networking platform, which provides multiple digital services to its users, has accessed a digital service of the social networking platform less than a threshold amount and (2) in response to determining that the user has accessed the digital service less than the threshold amount, presenting a selectable tile, which serves as an entry point for the digital service, in a digital tray of entry point tiles provided via a social media feed configured for the user. Various other methods, systems, and computer-readable media are also disclosed.
G06F 16/9535 - Adaptation de la recherche basée sur les profils des utilisateurs et la personnalisation
G06F 16/951 - IndexationTechniques d’exploration du Web
G06Q 50/00 - Technologies de l’information et de la communication [TIC] spécialement adaptées à la mise en œuvre des procédés d’affaires d’un secteur particulier d’activité économique, p. ex. aux services d’utilité publique ou au tourisme
A method for generating a realistic avatar of a user by an electronic device is presented. The method includes displaying, by at least one display, a sequence of frames to a user of the electronic device. The sequence of frames includes a first frame including a first color component, a second frame including a second color component, and third frame including a third color component. while displaying the sequence of images to the user, the method includes capturing, by one or more cameras, a plurality of images of the user. The method further includes determining, for each of the plurality of images, a visible wavelength for each of the first color component, the second color component, and the third color component, and generating a realistic avatar of the user based on the visible wavelength for each of the first color component, the second color component, and the third color component.
A61B 5/00 - Mesure servant à établir un diagnostic Identification des individus
A61B 5/0205 - Évaluation simultanée de l'état cardio-vasculaire et de l'état d'autres parties du corps, p. ex. de l'état cardiaque et respiratoire
A61B 5/1455 - Mesure des caractéristiques du sang in vivo, p. ex. de la concentration des gaz dans le sang ou de la valeur du pH du sang en utilisant des capteurs optiques, p. ex. des oxymètres à photométrie spectrale
G06T 7/90 - Détermination de caractéristiques de couleur
31.
ARTIFICIAL REALITY DEVICES WITH LIGHT BLOCKING CAPABILITY AND PROJECTION OF VISUAL CONTENT OVER REGIONS OF BLOCKED LIGHT
A head-mounted display includes a filter that at least partially blocks light emitted from a light source. In some examples, the light source may be a liquid crystal display that emits polarized light, and the filter may be a polarizing component. Using the filter, the head-mounted display may filter out the light source directed to the head-mounted display. The filtered out region may appear as a grayscale region or a blacked out region mixed with a real-world environment. When the head-mounted display takes the form of an augmented reality device, the head-mounted display may use a display to project visual content onto the filtered out region, while permitting a view of a real-world environment. The visual content may include images, video(s), and/or text, with the images, video(s), and/or text being relevant to the user.
The present application describes systems, methods, devices, and computer program products for convolutional neural networks (CNN) applicable for image processing, image scaling, and computer vision-oriented operations. Various embodiments for image scaling may receive image data corresponding to a first resolution. The image data may have a channel size and a data size. A CNN may be applied to process the image data according to a set of kernels. A first kernel set and a second kernel set may be independently applied to the image data to generate a first output set and a second output set. An interleaved set may be generated from the first output set and the second output set. An output image having a second data size may be generated from the output sets.
The present application at least describes a method including a step of receiving, at a convolutional neural network (CNN), data over a network from a source. The CNN may include one or more blocks. Each block may include plural layers. The method may include a step of causing, via the CNN in a first layer of the first block, a representation of the received data as a first matrix having M rows and N columns. The M rows and N columns may be greater than or equal to 1. The method may also include a step of processing, via the CNN at the first layer of the first block, the first matrix via a predetermined kernel matrix. The kernel matrix may include M-X rows and N-Y columns. X and Y may be greater than or equal to 1. The method may also include a step of rendering, via the CNN based on the processed first matrix, a second matrix having M-2 rows and N-2 columns. The method may further include a step of causing, via the CNN in a second layer of the first block, a representation including a first buffer and the second matrix. The first buffer may include at least 2 columns of the first matrix. The method may include yet a further step of processing, via the CNN at the second layer of the first block, the second matrix via the predetermined kernel matrix. The method may include yet even a further step of rendering, via the CNN based on the processed second matrix, a third matrix having M-4 rows and N-4 columns.
In particular embodiments, a computing system may receive a video comprising a plurality of image frames. The system may generate, for each image frame in the video, an initial depth map using a machine-learning model. The system may compute a misalignment error indicating depth misalignments in initial depth maps using a reprojection technique. The system may generate, for each image frame in the video, an optimized camera pose and a flexible deformation spline associated with the image frame to minimize the misalignment error. The system may generate, for each image frame in the video, a refined depth map by adjusting the initial depth map associated with the frame using the flexible deformation spline associated with the image frame.
In one embodiment, a method includes receiving touch inputs from a user corresponding to an activation trigger for an assistant system executing on a head-mounted device at the head-mounted device, accessing signals from inertial measurement unit (IMU) sensors of the head-mounted device by the head-mounted device, determining that the user is either donning or doffing the head-mounted device by an on-device don/doff detection model and based only on the signals from the IMU sensors, and overriding the activation trigger to prevent an activation of the assistant system responsive to the received touch inputs.
In one embodiment, a method includes receiving at a head-mounted device a speech input from a user and a visual input captured by cameras of the head-mounted device, wherein the visual input comprises subjects and attributes associated with the subjects, and wherein the speech input comprises a co-reference to one or more of the subjects, resolving entities corresponding to the subjects associated with the co-reference based on the attributes and the co-reference, and presenting a communication content responsive to the speech input and the visual input at the head-mounted device, wherein the communication content comprises information associated with executing results of tasks corresponding to the resolved entities.
G06Q 50/00 - Technologies de l’information et de la communication [TIC] spécialement adaptées à la mise en œuvre des procédés d’affaires d’un secteur particulier d’activité économique, p. ex. aux services d’utilité publique ou au tourisme
G06F 7/14 - Interclassement, c.-à-d. association d'au moins deux séries de supports d'enregistrement, chacun étant rangé dans le même ordre de succession, en vue de former une série unique rangée dans le même ordre de succession
G06F 9/451 - Dispositions d’exécution pour interfaces utilisateur
G06F 16/176 - Support d’accès partagé aux fichiersSupport de partage de fichiers
G06F 16/22 - IndexationStructures de données à cet effetStructures de stockage
G06F 16/951 - IndexationTechniques d’exploration du Web
G06F 16/9535 - Adaptation de la recherche basée sur les profils des utilisateurs et la personnalisation
G06F 18/2411 - Techniques de classification relatives au modèle de classification, p. ex. approches paramétriques ou non paramétriques basées sur la proximité d’une surface de décision, p. ex. machines à vecteurs de support
G06F 40/40 - Traitement ou traduction du langage naturel
G06N 3/006 - Vie artificielle, c.-à-d. agencements informatiques simulant la vie fondés sur des formes de vie individuelles ou collectives simulées et virtuelles, p. ex. simulations sociales ou optimisation par essaims particulaires [PSO]
G06V 10/82 - Dispositions pour la reconnaissance ou la compréhension d’images ou de vidéos utilisant la reconnaissance de formes ou l’apprentissage automatique utilisant les réseaux neuronaux
G06V 20/30 - ScènesÉléments spécifiques à la scène dans les albums, les collections ou les contenus partagés, p. ex. des photos ou des vidéos issus des réseaux sociaux
G06V 40/16 - Visages humains, p. ex. parties du visage, croquis ou expressions
G06V 40/20 - Mouvements ou comportement, p. ex. reconnaissance des gestes
G10L 13/00 - Synthèse de la paroleSystèmes de synthèse de la parole à partir de texte
G10L 13/04 - Détails des systèmes de synthèse de la parole, p. ex. structure du synthétiseur ou gestion de la mémoire
G10L 15/02 - Extraction de caractéristiques pour la reconnaissance de la paroleSélection d'unités de reconnaissance
G10L 15/06 - Création de gabarits de référenceEntraînement des systèmes de reconnaissance de la parole, p. ex. adaptation aux caractéristiques de la voix du locuteur
H04L 12/28 - Réseaux de données à commutation caractérisés par la configuration des liaisons, p. ex. réseaux locaux [LAN Local Area Networks] ou réseaux étendus [WAN Wide Area Networks]
H04L 41/00 - Dispositions pour la maintenance, l’administration ou la gestion des réseaux de commutation de données, p. ex. des réseaux de commutation de paquets
H04L 41/22 - Dispositions pour la maintenance, l’administration ou la gestion des réseaux de commutation de données, p. ex. des réseaux de commutation de paquets comprenant des interfaces utilisateur graphiques spécialement adaptées [GUI]
H04L 43/0882 - Utilisation de la capacité de la liaison
H04L 51/02 - Messagerie d'utilisateur à utilisateur dans des réseaux à commutation de paquets, transmise selon des protocoles de stockage et de retransmission ou en temps réel, p. ex. courriel en utilisant des réactions automatiques ou la délégation par l’utilisateur, p. ex. des réponses automatiques ou des messages générés par un agent conversationnel
H04L 51/046 - Interopérabilité avec d'autres applications ou services réseau
H04L 51/216 - Gestion de l'historique des conversations, p. ex. regroupement de messages dans des sessions ou des fils de conversation
H04L 67/10 - Protocoles dans lesquels une application est distribuée parmi les nœuds du réseau
H04L 67/53 - Services réseau en utilisant des fournisseurs tiers de services
H04L 67/5651 - Conversion ou adaptation du format ou du contenu d'applications en réduisant la quantité ou la taille des données d'application échangées
H04L 67/75 - Services réseau en affichant sur l'écran de l'utilisateur les conditions du réseau ou d'utilisation
In one embodiment, a method includes accessing a first document, accessing a plurality of second documents, calculating a relevance score for each of the plurality of second documents indicating a degree of relevance of the second document to the first document using an encoder of a machine-learning model, selecting a subset of the second documents based on their corresponding relevance scores, generating a target document by using the machine-learning model to process the subset of second documents and their corresponding relevance scores, and updating parameters of the machine-learning model based on a comparison between the first document and the generated target document.
G06F 40/58 - Utilisation de traduction automatisée, p. ex. pour recherches multilingues, pour fournir aux dispositifs clients une traduction effectuée par le serveur ou pour la traduction en temps réel
G06F 18/22 - Critères d'appariement, p. ex. mesures de proximité
The disclosed systems for suspending cable (e.g., fiber optic cable) from an overhead powerline may include a payload subsystem for housing and dispensing a cable along an overhead powerline, a rotation subsystem for winging the cable from the payload subsystem around the powerline, an extension subsystem for moving the payload subsystem to avoid obstacles, an obstacle detection subsystem for automatically detecting obstacles encountered along the powerline, a drive subsystem for driving the system along the powerline, and at least one processor for controlling the payload subsystem, rotation subsystem, extension subsystem, obstacle detection subsystem, and drive subsystem in a manner that avoids obstacles as the system moves along the powerline. Various other related systems, devices, components, and methods are also disclosed.
G06T 7/521 - Récupération de la profondeur ou de la forme à partir de la télémétrie laser, p. ex. par interférométrieRécupération de la profondeur ou de la forme à partir de la projection de lumière structurée
G06T 7/593 - Récupération de la profondeur ou de la forme à partir de plusieurs images à partir d’images stéréo
G06T 7/70 - Détermination de la position ou de l'orientation des objets ou des caméras
H02G 1/04 - Méthodes ou appareils spécialement adaptés à l'installation, entretien, réparation, ou démontage des câbles ou lignes électriques pour lignes ou câbles aériens pour les monter ou les tendre
H02G 7/02 - Dispositifs pour ajustage ou maintien de la tension mécanique, p. ex. organe de compensation
H04N 13/239 - Générateurs de signaux d’images utilisant des caméras à images stéréoscopiques utilisant deux capteurs d’images 2D dont la position relative est égale ou en correspondance à l’intervalle oculaire
Methods, systems, and apparatuses for disguising or otherwise hiding big data access patterns and frequencies may include a replica being directly obtained from a key and a replica index, based on applying a pseudorandom permutation or pseudorandom function. Data object access frequencies may be quantized so that the ratios of quantized data object access frequencies are rational numbers of integers, which may allow for a uniform distribution of data object accesses. Fake accesses may be implemented by simulating an access schedule produced by a cryptographic primitive and by performing a greedy mapping between the real data object accesses and simulated data object accesses.
A method implemented by a computing device includes displaying on a display of the computing device an extended reality (XR) environment, and determining one or more virtual characteristics associated with a first virtual content and a second visual content viewable within the displayed XR environment, in which the second virtual content is at least partially occluded by the first virtual content. The method further includes generating, based on the one or more virtual characteristics, a plurality of user input interception layers to be associated with the first virtual content and the second visual content, and in response to determining a user intent to interact with the second virtual content, directing one or more user inputs to the second virtual content based on whether or not the one or more user inputs are intercepted by one or more of the plurality of user input interception layers.
According to examples, a system for implementing image modification functions via use of variable scanning orders is described. The system may include a processor and a memory storing instructions. The processor, when executing the instructions, may cause the system to partition an image into a plurality of image blocks, identify one or more image blocks of the plurality of image blocks associated with a region of interest (ROI), and scan the one or more image blocks in an image modification order. The processor, when executing the instructions, may then arrange the one or more image blocks according to the image modification order to form a modified image including the region of interest (ROI) and crop the region of interest (ROI) in the modified image to form a new image.
A method includes receiving, from a client system, one or more utterances comprising one or more first words in a first language and one or more second words in a second language. The method further includes generating, based on a single bilingual automatic-speech-recognition (ASR) model, a transcription of the one or more utterances, such that the transcription comprises one or more first text strings in the first language and one or more second text strings in the second language. The method further includes executing one or more tasks based on the one or more first text strings in the first language and the one or more second text strings in the second language, and sending, to the client system, instructions for presenting a response responsive to the one or more utterances, such that the response is based on both the first and second languages.
G06F 40/58 - Utilisation de traduction automatisée, p. ex. pour recherches multilingues, pour fournir aux dispositifs clients une traduction effectuée par le serveur ou pour la traduction en temps réel
G10L 15/183 - Classement ou recherche de la parole utilisant une modélisation du langage naturel selon les contextes, p. ex. modèles de langage
G10L 15/30 - Reconnaissance distribuée, p. ex. dans les systèmes client-serveur, pour les applications en téléphonie mobile ou réseaux
43.
Large Language Models for Voice-Driven NPC Interactions
In one embodiment, a method includes receiving, by a mixed reality (MR) display device, an audio input from a first user of the MR display device, where the MR display device is associated with an MR environment including several MR objects, processing, using a natural language understanding (NLU) model, the audio input to identify one or more intents and one or more slots associated with the audio input, identifying a first MR object from several MR objects that is in an active listening state, where the first MR object is associated with a first set of intents and a first set of slots, determining that either the first set of intents or the first set of slots do not include the one or more identified intents or the one or more identified slots associated with the audio input, and generating, using a large language model (LLM), an out-of-domain (OOD) response.
In one embodiment, a method includes accessing from a client system associated with a first user sensor signals captured by sensors of the client system, wherein the client system comprises a plurality of sensors, and wherein the sensors signals are accessed from the sensors based on cascading model policies, wherein each cascading model policy utilizes one or more of a respective cost or relevance associated with each sensor, detecting a change in a context of the first user associated with an activity of the first user based on machine-learning models and the sensor signals, wherein the change in the context of the first user satisfies a trigger condition associated with the activity, and responsive to the detected change in the context of the first user automatically capturing visual data by cameras of the client system.
In one embodiment, a method includes receiving a voice input having acoustic features from a first client system associated with a first user, determining emotions associated with the voice input based on one or more of the acoustic features by machine-learning models, determining facial features for a first extended-reality (XR) avatar representing the first user based on the emotions, and sending instructions for rendering the first XR avatar representing the first user to a second client system associated with a second user, wherein the first XR avatar is rendered with the determined facial features.
G06T 13/40 - Animation tridimensionnelle [3D] de personnages, p. ex. d’êtres humains, d’animaux ou d’êtres virtuels
G10L 25/63 - Techniques d'analyse de la parole ou de la voix qui ne se limitent pas à un seul des groupes spécialement adaptées pour un usage particulier pour comparaison ou différentiation pour estimer un état émotionnel
Three-dimensional chip architecture is described herein. In one example aspect, an integrated circuit may include an interposer layer. The integrated circuit may further include a plurality of random access memory chiplets stacked atop the interposer layer, and a plurality of compute chiplets. The plurality of compute chiplets may be stacked atop a respective random access memory chip of the plurality of random access memory chiplets, such that the plurality of compute chiplets may be in electrical communication with the respective random access memory chip of the plurality of random access memory chiplets.
H10B 80/00 - Ensembles de plusieurs dispositifs comprenant au moins un dispositif de mémoire couvert par la présente sous-classe
H01L 25/16 - Ensembles consistant en une pluralité de dispositifs à semi-conducteurs ou d'autres dispositifs à l'état solide les dispositifs étant de types couverts par plusieurs des sous-classes , , , , ou , p. ex. circuit hybrides
48.
METHODS, APPARATUSES AND COMPUTER PROGRAM PRODUCTS FOR DETERMINING OPTIMIZED PERSONALIZED EXECUTION TIMES FOR DELIVERY OF OPTIMIZED CONTENT
A system for determining personalized execution times for delivery of optimized content is disclosed. The system may evaluate activities of a user, occurring within a network, during hours of days. The system may determine, based on the evaluated activities, optimal time periods during a time interval in which to schedule provision of content items to a communication device(s) associated with the user. The system may determine a best time period among the optimal time periods. The system may generate, in advance of the optimal time periods, content items tailored to the user. The system may determine, in advance of the best time period, a highest ranked generated content item, among the generated content items, for delivery to the communication device(s) associated with the user during the best time period. The system may enable provision, to the communication device(s), of the highest ranked generated content item during the best time period.
An online system receives explicit user data and explicit event data, and implicit user data and implicit event data from a third party system. The online system generates an implicit users/implicit events data feature, an explicit users/explicit events data feature, and an explicit users/implicit events data feature. The online system generates a prediction of the counterfactual rate based on the implicit users/implicit events data feature, the explicit users/explicit events data feature, and the explicit users/explicit events data feature, the counterfactual rate indicating the likelihood that target users matching certain characteristics caused an event to occur when the target are not been presented with content by the online system, the content configured to induce users to cause the event to occur. A combined prediction rate is presented to the third party system based on the counterfactual rate.
G06N 3/04 - Architecture, p. ex. topologie d'interconnexion
G06Q 30/02 - MarketingEstimation ou détermination des prixCollecte de fonds
G06Q 50/00 - Technologies de l’information et de la communication [TIC] spécialement adaptées à la mise en œuvre des procédés d’affaires d’un secteur particulier d’activité économique, p. ex. aux services d’utilité publique ou au tourisme
50.
Multimodal state tracking via scene graphs for assistant systems
In one embodiment, a method includes receiving, from a client system associated with a user, a first user request that includes a reference to a target object and one or more of an attribute or a relationship of the target object. Visual data including one or more images portraying the target object may then be accessed, and the reference may be resolved to the target object portrayed in the one or more images. Object information of the target object that corresponds to the referenced attribute or relationship of the first user request may be determined based on a visual analysis of the one or more images. Finally, responsive to receiving the first user request, the object information of the target object may be stored in a multimodal dialog state.
Systems and methods for social media video streaming may include (1) presenting, within a short-form video being streamed as part of a short-form video feed, a selectable theme element, indicating a theme associated with the short-form video, (2) receiving user input selecting the selectable theme element, and (3) in response to receiving the user input, presenting a digest interface of short-form videos, relating to the theme, which includes (i) sets of tiles, in which each tile corresponds to a different short-form video, organized into subthemes, and (ii) a digital social engagement element, which provides an entry point for digital social engagement relating to the theme.
H04N 21/482 - Interface pour utilisateurs finaux pour la sélection de programmes
H04L 65/61 - Diffusion en flux de paquets multimédias pour la prise en charge des services de diffusion par flux unidirectionnel, p. ex. radio sur Internet
52.
System and method for positioning a 3D eyeglasses model
In some embodiments, a computer-implemented method includes obtaining a without-eyeglasses face scan of a subject, the without-eyeglasses face scan being a three-dimensional (3D) model of a face of the subject without eyeglasses; obtaining a with-eyeglasses face scan of the subject, the with-eyeglasses face scan being a 3D model of the subject with eyeglasses; and using the without-eyeglasses face scan and the with-eyeglasses face scan to place a 3D eyeglasses model on a face model of the subject. In some embodiments of the computer-implemented method, the 3D eyeglasses model is placed on the face model of the subject using frame placement information generated using the without-eyeglasses face scan and the with-eyeglasses face scan.
G06K 9/00 - Méthodes ou dispositions pour la lecture ou la reconnaissance de caractères imprimés ou écrits ou pour la reconnaissance de formes, p.ex. d'empreintes digitales
G06T 7/73 - Détermination de la position ou de l'orientation des objets ou des caméras utilisant des procédés basés sur les caractéristiques
G06T 17/10 - Description de volumes, p. ex. de cylindres, de cubes ou utilisant la GSC [géométrie solide constructive]
G06T 19/20 - Édition d'images tridimensionnelles [3D], p. ex. modification de formes ou de couleurs, alignement d'objets ou positionnements de parties
Methods, apparatuses, and systems for using a compensating window to correct tolerance-placement effects on camera focus are provided. The system may receive a first captured image of a first test target from a surface of a target plane. The first captured image may be captured using a first lens of a camera. The system may determine a first modulation transfer function measurement for the first captured image. The system may determine that the first modulation transfer function measurement is within a threshold measurement. The system may send an alert indicative that the first lens is within the threshold measurement.
G06F 3/01 - Dispositions d'entrée ou dispositions d'entrée et de sortie combinées pour l'interaction entre l'utilisateur et le calculateur
H04N 23/56 - Caméras ou modules de caméras comprenant des capteurs d'images électroniquesLeur commande munis de moyens d'éclairage
H04N 23/60 - Commande des caméras ou des modules de caméras
H04N 25/615 - Traitement du bruit, p. ex. détection, correction, réduction ou élimination du bruit le bruit provenant uniquement de l'objectif, p. ex. l'éblouissement, l'ombrage, le vignettage ou le "cos4" impliquant une fonction de transfert modélisant le système optique, p. ex. une fonction de transfert optique [OTF], une fonction de transfert de phase [PhTF] ou une fonction de transfert de modulation [MTF]
55.
Display screen with an animated graphical user interface
Technology for customized crest factor reduction (CFR) noise shaping includes dividing a frequency band into a plurality of regions, assigning a constellation goal for each region, the respective constellation goal for at least two regions being different, determining a CFR noise level for each region based on the constellation goal for the region and a target CFR noise level for the divided frequency band, creating a cancellation pulse based on scaling factors, and based on the cancellation pulse, applying a cancellation pulse signal on a per-region basis to generate transmission signals having the determined CFR noise level for each region. In examples, a first region has a first constellation goal and a second region has a second constellation goal, and a determined CFR noise level for the first region supports the first constellation goal and a determined CFR noise level for the second region supports the second constellation goal.
H04L 27/26 - Systèmes utilisant des codes à fréquences multiples
H04L 27/34 - Systèmes à courant porteur à modulation de phase et d'amplitude, p. ex. en quadrature d'amplitude
57.
Methods, apparatuses and computer program products for providing transmission chirped volume bragg grating based compact waveguide in-couplers for light sources
Methods and systems for propagating light into a waveguide are provided. The system may include a light source configured to generate light. The system may include at least one mirror configured to direct the light into one or more rays of light. The system may include a Surface Relief Grating disposed on a Volume Bragg Grating. The Surface Relief Grating may receive the one or more rays of light and may diffract the one or more rays of light. The Volume Bragg Grating may be disposed on the waveguide in which the waveguide may be configured to receive the one or more rays of light from the Volume Bragg Grating and propagate the one or more rays of light throughout the waveguide such that an off-Bragg condition is exhibited by the one or more rays of light propagating through the waveguide.
G09G 5/00 - Dispositions ou circuits de commande de l'affichage communs à l'affichage utilisant des tubes à rayons cathodiques et à l'affichage utilisant d'autres moyens de visualisation
G03H 1/02 - Procédés ou appareils holographiques utilisant la lumière, les infrarouges ou les ultraviolets pour obtenir des hologrammes ou pour en obtenir une imageLeurs détails spécifiques Détails
58.
Systems and methods for incorporating avatars into real-time communication sessions
The disclosed computer-implemented method may include systems for incorporating a user's avatar into a real-time communication session. For example, the described systems establish a real-time communication session between two or more social networking system users. The described systems further generate a landmark map representing positioning of one of the real-time communication session participants, and transmit the landmark map with the participant's avatar to one or more recipients. On the recipient-side, the described systems render the transmitted avatar according to the landmark map. Various other methods, systems, and computer-readable media are also disclosed.
A head-mounted display including multiple displays tiled together is provided. A first display may include an active region and an inactive region. A second display may be aligned with the inactive region. The first display and the second display may have different resolutions and may present visual information with different display qualities based on their respective resolutions. For example, the first display may include a lower resolution than the second display. The inactive region may be centered in the first display, and the second display may be centered with respect to the first display. The head-mounted display may provide a continuous image(s) (or video(s)) with a high resolution on the second display, while providing the remainder of the image(s) on the low resolution first display. Based on the multiple displays being tiled, the multiple displays may present one or more images, or one or more videos, in a non-overlapping manner.
G09G 3/3208 - Dispositions ou circuits de commande présentant un intérêt uniquement pour l'affichage utilisant des moyens de visualisation autres que les tubes à rayons cathodiques pour la présentation d'un ensemble de plusieurs caractères, p. ex. d'une page, en composant l'ensemble par combinaison d'éléments individuels disposés en matrice utilisant des sources lumineuses commandées utilisant des panneaux électroluminescents semi-conducteurs, p. ex. utilisant des diodes électroluminescentes [LED] organiques, p. ex. utilisant des diodes électroluminescentes organiques [OLED]
60.
CHIP-TO-CHIP INTERCONNECT WITH A LAYERED COMMUNICATION ARCHITECTURE
A system includes a first integrated circuit package including a first group of one or more artificial intelligence processing units and a first chip-to-chip interconnect communication unit and a second integrated circuit package including a second group of one or more artificial intelligence processing units and a second chip-to-chip interconnect communication unit. The system also includes an interconnect between the first integrated circuit package and the second integrated circuit package, wherein the first chip-to-chip interconnect communication unit and the second chip-to-chip interconnect communication unit manage Ethernet-based communication via the interconnect using a layered communication architecture supporting a credit-based data flow control and a retransmission data flow control.
In one embodiment, a method includes receiving a first user request from a first user for generating a media montage from a client system during a dialog session with the first user, generating an initial media montage during the dialog session based on media collections associated with the first user, sending instructions for presenting the initial media montage to the client system during the dialog session, receiving a second user request from the first user from the client system during the dialog session for editing the initial media montage, generating an edited media montage from the initial media montage during the dialog session based on the second user request and a memory graph associated with the first user, and sending instructions for presenting the edited media montage to the client system during the dialog session.
The disclosed computer-implemented method may include systems and methods for embedding specific data into a call stack associated with an application session. For example, the systems and methods described herein can initialize a program thread that sequentially executes specialized application functions based on characters of a unique identifier to embed the unique identifier within a call stack of the application session. The systems and methods further provide the unique identifier in connection with other data sources associated with the application session such that further analysis of all data associated with the application session may be cross-referenced according to the unique identifier. Various other methods, systems, and computer-readable media are also disclosed.
The disclosed systems and methods may include (1) providing a selectable dashboard element within an interface of a dating application configured for a user of the dating application, (2) receiving user input selecting the dashboard element, and (3) in response to receiving the user input, presenting a digital dating-analytics dashboard that provides information relating to the user's dating history aggregated via the dating application. Various other methods, systems, and computer-readable media are also disclosed.
G06Q 50/00 - Technologies de l’information et de la communication [TIC] spécialement adaptées à la mise en œuvre des procédés d’affaires d’un secteur particulier d’activité économique, p. ex. aux services d’utilité publique ou au tourisme
G06F 3/04842 - Sélection des objets affichés ou des éléments de texte affichés
G06F 16/9535 - Adaptation de la recherche basée sur les profils des utilisateurs et la personnalisation
The disclosed systems and methods may include (1) receiving, from a user who has a dating account with a dating application maintained by a social media platform, user input designating a contact of the user, connected to the user via the social media platform, as a matchmaker for the dating account and (2) in response to receiving the user input designating the contact as a matchmaker for the dating account, providing the contact with access to a recommended-matches interface of the dating application, comprising a queue of recommended dating matches selected for the user, in a matchmaker mode that enables the contact to perform one or more digital matchmaking actions. Various other methods, systems, and computer-readable media are also disclosed.
G06Q 50/00 - Technologies de l’information et de la communication [TIC] spécialement adaptées à la mise en œuvre des procédés d’affaires d’un secteur particulier d’activité économique, p. ex. aux services d’utilité publique ou au tourisme
66.
Private language model adaptation for speech recognition
In one embodiment, a method includes accessing a decoded hypothesis corresponding to an utterance, computing a predicted probability of observing each token in the decoded hypothesis by having a local first machine-learning model process the decoded hypothesis, computing a confidence score for each token in the decoded hypothesis by having a second machine-learning model process the decoded hypothesis, where the confidence score indicates a degree of confidence for the token to be observed at its position, calculating a loss for the computed predicted probabilities of observing tokens in the decoded hypothesis based on the computed confidence scores, and updating parameters of the local first machine-learning model based on the calculated loss.
G10L 15/16 - Classement ou recherche de la parole utilisant des réseaux neuronaux artificiels
G10L 15/06 - Création de gabarits de référenceEntraînement des systèmes de reconnaissance de la parole, p. ex. adaptation aux caractéristiques de la voix du locuteur
G10L 15/30 - Reconnaissance distribuée, p. ex. dans les systèmes client-serveur, pour les applications en téléphonie mobile ou réseaux
G10L 15/32 - Reconnaisseurs multiples utilisés en séquence ou en parallèleSystèmes de combinaison de score à cet effet, p. ex. systèmes de vote
67.
Representing Real-World Objects with a Virtual Reality Environment
An image processing system enables a user wearing a head-mounted display to experience a virtual environment combined with a representation of a real-world object. The image processing system receives a captured scene of a real-world environment that includes a target object. The image processing system identifies the target object in the captured scene and generates a representation of the target object. In some cases, the image processing system may include a graphical overlay with the representation of the target object. The image processing system can generate a combined scene that includes the target object and the virtual environment. The combined scene is presented to the user, thereby allowing the user to interact with the real-world target object (or a representation thereof) in combination with the virtual environment.
G06F 3/00 - Dispositions d'entrée pour le transfert de données destinées à être traitées sous une forme maniable par le calculateurDispositions de sortie pour le transfert de données de l'unité de traitement à l'unité de sortie, p. ex. dispositions d'interface
G06T 19/00 - Transformation de modèles ou d'images tridimensionnels [3D] pour infographie
68.
Display screen or portion thereof with a graphical user interface
Systems, apparatuses and methods provide technology that compresses first data based on a first compression scheme to generate second data, where the first data is associated with a first machine learning model. The technology stores the second data into a memory, adjusts a first entry of a lookup table to correspond to the first compression scheme based on the first data being compressed based on the first compression scheme, provide the second data from the memory to processing elements of a processing array during execution of the first machine learning model, and decompresses, at the processing array, the second data based on the lookup table to obtain the first data.
H03M 7/42 - Conversion en, ou à partir de codes de longueur variable, p. ex. code Shannon-Fano, code Huffman, code Morse utilisant une table pour le procédé de codage ou de décodage, p. ex. utilisant une mémoire morte
71.
Display screen with an animated graphical user interface
Various aspects of the subject technology relate to systems, methods, and machine-readable media for bundling data into media files. The disclosed method comprises receiving a player video file, wherein the player video file is generated on a social media platform. The method includes identifying a storage section of the player video file configured to store an editable video file. The method includes generating the editable video file comprising metadata that includes a plurality of parameters associated with the player video file. A parameter of the plurality of parameters identifies an editable element of the player video file. The method includes generating a bundled video file by appending the editable video file in the storage section of the player video file. The method comprises exporting the bundled video file to storage located external to the social media platform.
H04N 21/231 - Opération de stockage de contenu, p. ex. mise en mémoire cache de films pour stockage à court terme, réplication de données sur plusieurs serveurs, ou établissement de priorité des données pour l'effacement
G06Q 50/00 - Technologies de l’information et de la communication [TIC] spécialement adaptées à la mise en œuvre des procédés d’affaires d’un secteur particulier d’activité économique, p. ex. aux services d’utilité publique ou au tourisme
73.
PRECISION TIME PROTOCOL REDUNDANCY AND LOAD BALANCING ACROSS A SET OF ACTIVE TIME SERVERS
A system for achieving Precision Time Protocol (PTP) redundancy and load balancing across a set of active time servers deployed in a cluster is disclosed. The system may organize a set of PTP Grandmasters (GMs) (e.g., located in a datacenter) into a plurality of PTP clusters (e.g., for load balancing). A set of PTP ordinary clocks (OCs) (e.g., time clients) may be grouped in OC Groups and then into PTP clusters for GM assignation, wherein each PTP OC includes a unicast master table (UMT) that includes one or more potential PTP GM port addresses and priority values. Based on a Master Clock Mechanism, one of the potential GMs may be selected based on one or more PTP GM priority attributes.
In one embodiment, a method includes receiving, at a client system, an audio input, where the audio input comprises a coreference to a target object, accessing visual data from one or more camera associated with the client system, where the visual data comprises images portraying one or more objects, resolving the coreference to the target object from among the one or more objects, resoling the target object to a specific entity, and providing, at the client system, a response to the audio input, where the response comprises information about the specific entity.
G06Q 50/00 - Technologies de l’information et de la communication [TIC] spécialement adaptées à la mise en œuvre des procédés d’affaires d’un secteur particulier d’activité économique, p. ex. aux services d’utilité publique ou au tourisme
G06F 3/01 - Dispositions d'entrée ou dispositions d'entrée et de sortie combinées pour l'interaction entre l'utilisateur et le calculateur
G06F 16/9536 - Personnalisation de la recherche basée sur le filtrage social ou collaboratif
G06F 18/2321 - Techniques non hiérarchiques en utilisant les statistiques ou l'optimisation des fonctions, p. ex. modélisation des fonctions de densité de probabilité
G06V 10/764 - Dispositions pour la reconnaissance ou la compréhension d’images ou de vidéos utilisant la reconnaissance de formes ou l’apprentissage automatique utilisant la classification, p. ex. des objets vidéo
G06V 10/82 - Dispositions pour la reconnaissance ou la compréhension d’images ou de vidéos utilisant la reconnaissance de formes ou l’apprentissage automatique utilisant les réseaux neuronaux
G06V 20/00 - ScènesÉléments spécifiques à la scène
G06V 20/20 - ScènesÉléments spécifiques à la scène dans les scènes de réalité augmentée
G06V 20/30 - ScènesÉléments spécifiques à la scène dans les albums, les collections ou les contenus partagés, p. ex. des photos ou des vidéos issus des réseaux sociaux
G06V 20/40 - ScènesÉléments spécifiques à la scène dans le contenu vidéo
G06V 40/16 - Visages humains, p. ex. parties du visage, croquis ou expressions
G06V 40/20 - Mouvements ou comportement, p. ex. reconnaissance des gestes
G10L 15/06 - Création de gabarits de référenceEntraînement des systèmes de reconnaissance de la parole, p. ex. adaptation aux caractéristiques de la voix du locuteur
H04L 51/212 - Surveillance ou traitement des messages utilisant un filtrage ou un blocage sélectif
H04L 51/222 - Surveillance ou traitement des messages en utilisant des informations de localisation géographique, p. ex. des messages transmis ou reçus à proximité d'un certain lieu ou d'une certaine zone
H04L 51/224 - Surveillance ou traitement des messages en fournissant une notification sur les messages entrants, p. ex. des poussées de notifications des messages reçus
H04L 51/52 - Messagerie d'utilisateur à utilisateur dans des réseaux à commutation de paquets, transmise selon des protocoles de stockage et de retransmission ou en temps réel, p. ex. courriel pour la prise en charge des services des réseaux sociaux
An image processing system enables a user wearing a head-mounted display to experience a virtual environment combined with a representation of a real-world object. The image processing system receives a captured scene of a real-world environment that includes a target object. The image processing system identifies the target object in the captured scene and generates a representation of the target object. In some cases, the image processing system may include a graphical overlay with the representation of the target object. The image processing system can generate a combined scene that includes the target object and the virtual environment. The combined scene is presented to the user, thereby allowing the user to interact with the real-world target object (or a representation thereof) in combination with the virtual environment.
G06F 3/00 - Dispositions d'entrée pour le transfert de données destinées à être traitées sous une forme maniable par le calculateurDispositions de sortie pour le transfert de données de l'unité de traitement à l'unité de sortie, p. ex. dispositions d'interface
G06T 19/00 - Transformation de modèles ou d'images tridimensionnels [3D] pour infographie
H04N 5/445 - Circuits de réception pour visualisation d'information additionnelle
78.
Generating Contextual Responses for Out-of-coverage Requests for Assistant Systems
In one embodiment, a method includes receiving a voice input from a first user from a client system via an assistant xbot associated with an assistant system, determining that the voice input includes an out-of-coverage request including intents or slots that are unsupported by the assistant system, generating a contextual response for the out-of-coverage request, wherein the contextual response includes an implicit confirmation of the out-of-coverage request and a failure response to the out-of-coverage request, wherein the implicit confirmation references the voice input from the first user, and wherein the failure response indicates that the assistant system is unable to execute a task corresponding to the out-of-coverage request, and sending instructions for presenting the contextual response responsive to the voice input to the client system.
Systems and methods for smart message delivery for message handoff between communication devices and artificial reality systems are provided. In various examples, a communication device or HMD may receive a message initiating a detection of motion(s) and/or position(s) of the communication device and/or the HMD associated with a user with respect to each other. The communication device and/or the HMD may determine a movement level associated with each other, and based on the movement level relative to a predetermined threshold a message delivery process may be determined. If the predetermined threshold is not met, the message may be output to a user via audio and if the predetermined threshold is met, text associated with the message may be presented by the communication device for the user to read. The system may further monitor the movement level while the message is being output to the user, and if the predetermined threshold is no longer met, the message may be output to the user via audio by the HMD based on where the user stopped reading text of the message. In situations where the movement level increases to meet or exceed the predetermined threshold, the message may be output via text by the communication device, and the text corresponding to the portion of the message read aloud as audio by the HMD may be in a different format to enable the user to easily determine where to start reading the message.
H04L 67/12 - Protocoles spécialement adaptés aux environnements propriétaires ou de mise en réseau pour un usage spécial, p. ex. les réseaux médicaux, les réseaux de capteurs, les réseaux dans les véhicules ou les réseaux de mesure à distance
80.
Systems and methods for screenless computerized social-media access
Systems and methods for screenless computerized social-media access may include (1) producing, via an audio speaker that is communicatively coupled to a computing device, a computer-generated verbal description of a social-media post provided via a social-media application, (2) detecting, via a microphone that is communicatively coupled to the computing device, an audible response to the social-media post from a user of the computing device, and (3) digitally responding to the social-media post in accordance with the detected audible response. Various other methods, systems, and computer-readable media are also disclosed.
G10L 15/22 - Procédures utilisées pendant le processus de reconnaissance de la parole, p. ex. dialogue homme-machine
G06Q 50/00 - Technologies de l’information et de la communication [TIC] spécialement adaptées à la mise en œuvre des procédés d’affaires d’un secteur particulier d’activité économique, p. ex. aux services d’utilité publique ou au tourisme
G10L 13/02 - Procédés d'élaboration de parole synthétiqueSynthétiseurs de parole
G10L 15/26 - Systèmes de synthèse de texte à partir de la parole
H04L 51/52 - Messagerie d'utilisateur à utilisateur dans des réseaux à commutation de paquets, transmise selon des protocoles de stockage et de retransmission ou en temps réel, p. ex. courriel pour la prise en charge des services des réseaux sociaux
81.
Systems and methods for prioritizing digital social media posts within a social media interface
The disclosed systems and methods may include (1) identifying social media posts that have been posted to a social media platform by a creator account with fewer than a designated number of followers, (2) creating a social-engagement score for the creator account based on a measure of an amount of digital social engagement generated in response to the social media posts, (3) determining that the social-engagement score surpasses a threshold, (4) in response to (i) the creator account having fewer than the designated number of followers and (ii) the social-engagement score surpassing the threshold, designating the creator account as a prioritized creator account, and (5) presenting, in association with the creator account, a digital visual indicator indicating that the creator account is a prioritized creator account. Various other methods, systems, and computer-readable media are also disclosed.
G06F 15/173 - Communication entre processeurs utilisant un réseau d'interconnexion, p. ex. matriciel, de réarrangement, pyramidal, en étoile ou ramifié
G06Q 50/00 - Technologies de l’information et de la communication [TIC] spécialement adaptées à la mise en œuvre des procédés d’affaires d’un secteur particulier d’activité économique, p. ex. aux services d’utilité publique ou au tourisme
Techniques for enabling users logged-out of a social networking system to access content associated with the social networking system are described. For example, based at least in part on determining a user is unassociated with a social networking system, the social networking system may present content to the user. In some examples, the amount of content available to the user may be limited, in which the user may be prompted to create an account. In some examples, logged-out users may be prompted to pick topics associated with their interests, allowing the social networking system to curate content for logged-out users.
G06F 16/958 - Organisation ou gestion de contenu de sites Web, p. ex. publication, conservation de pages ou liens automatiques
G06Q 50/00 - Technologies de l’information et de la communication [TIC] spécialement adaptées à la mise en œuvre des procédés d’affaires d’un secteur particulier d’activité économique, p. ex. aux services d’utilité publique ou au tourisme
83.
Preventing false activations based on don/doff detection for assistant systems
In one embodiment, a method includes receiving touch inputs from a user corresponding to an activation trigger for an assistant system executing on a head-mounted device at the head-mounted device, accessing signals from inertial measurement unit (IMU) sensors of the head-mounted device by the head-mounted device, determining that the user is either donning or doffing the head-mounted device by an on-device don/doff detection model and based only on the signals from the IMU sensors, and overriding the activation trigger to prevent an activation of the assistant system responsive to the received touch inputs.
The disclosed system may include (1) a drive subsystem that translates along a powerline conductor, (2) a rotation subsystem that rotates a segment of fiber optic cable about the powerline conductor while the drive subsystem translates along the powerline conductor such that the segment of fiber optic cable is wrapped helically about the powerline conductor, and (3) an extension subsystem that (a) mechanically couples the rotation subsystem to the drive subsystem, and (b) selectively extends the rotation subsystem away from the drive subsystem and the powerline conductor to avoid obstacles along the powerline conductor. Various other systems and methods are also disclosed.
H02G 1/02 - Méthodes ou appareils spécialement adaptés à l'installation, entretien, réparation, ou démontage des câbles ou lignes électriques pour lignes ou câbles aériens
The disclosed systems and methods may include (1) providing, within a digital help center article directed to delineating how to accomplish a digital action within an application, both (i) a description of how to manually accomplish the digital action (including a list of digital steps that, when manually initiated via user input to the application, trigger the digital action to be executed within the application) and (ii) a selectable element configured to trigger the digital action to be executed, bypassing one or more of the digital steps listed in the description, when selected via user input, (2) receiving user input digitally selecting the selectable element within the digital help center article, and (3) in response to receiving the user input digitally selecting the selectable element within the digital help center article, executing the digital action within the application. Various other methods, systems, and computer-readable media are also disclosed.
Methods, systems, and storage media for determining the similarities of post content for mapping into a hyperspace. In an exemplary method, the disclosure comprises receiving a query at the processor. The method includes determining post data associated with the query. The post data comprises a plurality of posts provided to a social media platform by various users of the platform. The method includes determining, by the processor, a relationship between at least two posts of the plurality of posts. The method includes training, by the processor, a machine language model. The machine language model is based on the query and the relationship between the at least two posts. The method also generates a hyperspace based on the relationship between the at least two posts and the query.
G06F 16/9536 - Personnalisation de la recherche basée sur le filtrage social ou collaboratif
G06Q 50/00 - Technologies de l’information et de la communication [TIC] spécialement adaptées à la mise en œuvre des procédés d’affaires d’un secteur particulier d’activité économique, p. ex. aux services d’utilité publique ou au tourisme
87.
DETECTING HEAD GESTURES USING INERTIAL MEASUREMENT UNIT SIGNALS
In one embodiment, a method includes presenting a suggestion to a user of a head-mounted device by the head-mounted device via an assistant xbot during a dialog session between the user and the assistant xbot, wherein the suggestion is associated with a plurality of actions to be performed by an assistant system associated with the assistant xbot, accessing signals from inertial measurement unit (IMU) sensors of the head-mounted device by the head-mounted device during the dialog session, determining a head gesture performed by the user during the dialog session by an on-device head-gesture detection model and based only on the signals from the IMU sensors, and executing a first action from multiple actions by the assistant system executing on the head-mounted device, wherein the first action is selected based on the determined head gesture during the dialog session.
G06F 3/01 - Dispositions d'entrée ou dispositions d'entrée et de sortie combinées pour l'interaction entre l'utilisateur et le calculateur
G06F 3/0346 - Dispositifs de pointage déplacés ou positionnés par l'utilisateurLeurs accessoires avec détection de l’orientation ou du mouvement libre du dispositif dans un espace en trois dimensions [3D], p. ex. souris 3D, dispositifs de pointage à six degrés de liberté [6-DOF] utilisant des capteurs gyroscopiques, accéléromètres ou d’inclinaison
Systems, methods, and non-transitory computer-readable media can determine at least one graphical overlay created by a content creator. The at least one graphical overlay can be distributed to at least one user through a social networking system. A determination can be made of a user selection of the at least one graphical overlay. The graphical overlay can be applied to at least one content item.
G06F 3/0482 - Interaction avec des listes d’éléments sélectionnables, p. ex. des menus
G06F 3/04845 - Techniques d’interaction fondées sur les interfaces utilisateur graphiques [GUI] pour la commande de fonctions ou d’opérations spécifiques, p. ex. sélection ou transformation d’un objet, d’une image ou d’un élément de texte affiché, détermination d’une valeur de paramètre ou sélection d’une plage de valeurs pour la transformation d’images, p. ex. glissement, rotation, agrandissement ou changement de couleur
H04L 51/52 - Messagerie d'utilisateur à utilisateur dans des réseaux à commutation de paquets, transmise selon des protocoles de stockage et de retransmission ou en temps réel, p. ex. courriel pour la prise en charge des services des réseaux sociaux
H04L 67/10 - Protocoles dans lesquels une application est distribuée parmi les nœuds du réseau
89.
Contextualized streaming end-to-end speech recognition with trie-based deep biasing and shallow fusion
In one embodiment, a method includes receiving a user's utterance comprising a word in a custom vocabulary list of the user, generating a previous token to represent a previous audio portion of the utterance, and generating a current token to represent a current audio portion of the utterance by generating a bias embedding by using the previous token to query a trie of wordpieces representing the custom vocabulary list, generating first probabilities of respective first candidate tokens likely uttered in the current audio portion based on the bias embedding and the current audio portion, generating second probabilities of respective second candidate tokens likely uttered after the previous token based on the previous token and the bias embedding, and generating the current token to represent the current audio portion of the utterance based on the first probabilities of the first candidate tokens and the second probabilities of the second candidate tokens.
In one embodiment, a method includes rendering a first output image comprising one or more augmented-reality (AR) objects for displays of an AR rendering device of an AR system associated with a first user. The method further includes accessing sensor signals associated with the first user. The one or more sensor signals may be captured by sensors of the AR system. The method further includes detecting a change in a context of the first user with respect to a real-world environment based on the sensor signals. The method further includes rendering a second output image comprising the AR objects for the displays of the AR rendering device. One or more of the AR objects may be adapted based on the detected change in the context of the first user.
G06F 1/3206 - Surveillance d’événements, de dispositifs ou de paramètres initiant un changement de mode d’alimentation
G06F 3/01 - Dispositions d'entrée ou dispositions d'entrée et de sortie combinées pour l'interaction entre l'utilisateur et le calculateur
G06F 3/04815 - Interaction s’effectuant dans un environnement basé sur des métaphores ou des objets avec un affichage tridimensionnel, p. ex. modification du point de vue de l’utilisateur par rapport à l’environnement ou l’objet
G06F 16/11 - Administration des systèmes de fichiers, p. ex. détails de l’archivage ou d’instantanés
G06T 19/00 - Transformation de modèles ou d'images tridimensionnels [3D] pour infographie
G06T 19/20 - Édition d'images tridimensionnelles [3D], p. ex. modification de formes ou de couleurs, alignement d'objets ou positionnements de parties
G10L 15/22 - Procédures utilisées pendant le processus de reconnaissance de la parole, p. ex. dialogue homme-machine
G10L 15/26 - Systèmes de synthèse de texte à partir de la parole
G10L 15/30 - Reconnaissance distribuée, p. ex. dans les systèmes client-serveur, pour les applications en téléphonie mobile ou réseaux
H04L 51/02 - Messagerie d'utilisateur à utilisateur dans des réseaux à commutation de paquets, transmise selon des protocoles de stockage et de retransmission ou en temps réel, p. ex. courriel en utilisant des réactions automatiques ou la délégation par l’utilisateur, p. ex. des réponses automatiques ou des messages générés par un agent conversationnel
Systems, methods, devices, and computer program products are provided for producing observable virtual images. Aspects may include at least one illumination source emitting light on a display and a transparent combining optic including a holographic optical element (HOE). According to various examples, light emitted from the at least one illumination source illuminates the transparent combining optic, and the transparent combining optic diffracts the light to generate an observable virtual image. The observable virtual image may be positioned to overlay a scene viewable through the transparent combining optic. Such aspects may be incorporated on a variety of technologies, such as head-mounted display systems, smart glasses, and/or AR devices.
A system and method for generating an eye box, an observable virtual image, and/or a multiplexed hologram are provided. The system may include a transparent combining optic including a holographic optical element. The transparent combining optic may be configured to diffract light received at a first side of the transparent combining optic. The light may be generated by an illumination source. The transparent combining optic may also be configured to form a virtual image viewable from a non-pupil-forming eyebox. The observable virtual image may be viewable from the first side of the transparent combining optic.
Systems, apparatuses and methods provide technology that identifies a plurality of data samples of a data set. In response to a trigger, the technology identifies a first subset of samples of the plurality of data samples that are outside a range associated with a decision boundary of a machine learning model, and removes the first subset of samples from the data set to generate a modified training set. The technology train the machine learning model based on the modified training set.
In one embodiment, a method for retrieving a compressed data chunk from a source memory to a data buffer using a direct-memory access includes generating a source address indicating a location in the source memory at which a metadata corresponding to a compressed data chunk is stored, reading the metadata from the source address, where the metadata includes a data address, a size and compression options associated with the compressed data chunk, reading the compressed data chunk from the source memory based on the data address and the size within the metadata, decompressing the compressed data chunk based on the compression options within the metadata, and storing the decompressed data chunk into the data buffer.
G06F 12/0802 - Adressage d’un niveau de mémoire dans lequel l’accès aux données ou aux blocs de données désirés nécessite des moyens d’adressage associatif, p. ex. mémoires cache
G06F 13/28 - Gestion de demandes d'interconnexion ou de transfert pour l'accès au bus d'entrée/sortie utilisant le transfert par rafale, p. ex. acces direct à la mémoire, vol de cycle
95.
Circular buffer for input and output of tensor computations
In one embodiment, a method includes receiving a token indicating a data chunk becomes available in a first circular buffer of a pre-determined size from a direct memory access component, determining that a computation is to be performed with data including the data chunk based on the token, and generating one or more addresses corresponding to one or more data chunks within the first circular buffer that are to be retrieved for the computation, where a generated address is subtracted by the pre-determined size of the first circular buffer when the generated address is greater than a pre-determined maximum associated with the first circular buffer, and where the generated address is added by the pre-determined size of the first circular buffer when the generated address is less than a pre-determined minimum associated with the first circular buffer.
G06F 13/28 - Gestion de demandes d'interconnexion ou de transfert pour l'accès au bus d'entrée/sortie utilisant le transfert par rafale, p. ex. acces direct à la mémoire, vol de cycle
Systems, methods, and non-transitory computer-readable media can be configured to perform operations comprising determining capabilities associated with a plurality of devices in a communication session, determining one or more media channels for transmission of media between the plurality of devices based on the capabilities, and facilitating the transmission of media between the plurality of devices via the one or more media channels.
Techniques are described for a social networking system to provide a collaborative fundraising event that allow for multiple organizers of the event to publish, receive attribution, and edit fundraiser content. In some cases, the social networking system may be configured to distribute the fundraiser content based on metrics, support, and relationships between organizers and other users of the social networking system.
G06Q 30/02 - MarketingEstimation ou détermination des prixCollecte de fonds
G06Q 50/00 - Technologies de l’information et de la communication [TIC] spécialement adaptées à la mise en œuvre des procédés d’affaires d’un secteur particulier d’activité économique, p. ex. aux services d’utilité publique ou au tourisme
H04L 51/52 - Messagerie d'utilisateur à utilisateur dans des réseaux à commutation de paquets, transmise selon des protocoles de stockage et de retransmission ou en temps réel, p. ex. courriel pour la prise en charge des services des réseaux sociaux
98.
Interaction Composer for Conversation Design Flow for Assistant Systems
In one embodiment, a method includes sending instructions for presenting a visual programming interface for a composer tool to a client system, wherein the visual programming interface comprises primitives for conversation design, wherein the primitives comprise at least an input-primitive, a response-primitive, and a decision-primitive, receiving instructions from a user for creating a conversation design flow for an application via the visual programming interface from the client system, wherein the conversation design flow comprises at least one or more input-primitives for one or more voice inputs and one or more input-primitives for one or more signal inputs, simulating an execution of the conversation design flow within the composer tool, and exporting the conversation design flow to a software package configured to be executable by the application, wherein the application is operable to process voice inputs and signal inputs according to the input-primitives of the conversation design flow.
In one embodiment, a method for accessing an instruction that is to be executed by a control agent within a computing system may include decompressing the instruction by replacing each of one or more zero-symbol run-length fields in the instruction with as many continuous zero symbols as its corresponding value and removing one or more non-zero-symbol run-length fields from the instruction. The method may also include determining that the instruction is spatial-delta-encoded based on a compression data header associated with the instruction, performing spatial-delta decoding on the instruction in response to the determination by orderly determining a spatial-delta-decoded value of each bit in the instruction, and causing the instruction to be sent to the control agent.
Systems and methods are provided for operating an artificial reality device using gestures. In various exemplary embodiments, an artificial reality device may receive information initiating a gesture recognition mode via the artificial reality device, and cause a camera associated with the artificial reality device to operate in the gesture recognition mode. The artificial reality device may further receive at least one image of a gesture captured by the camera and identify and evaluate the gesture. The gesture may be indicative of a user command. The artificial reality device may also execute the user command, and may generate feedback confirming execution of the user command.