SOUND FIELD ROTATION
Granted: May 8, 2025
Application Number:
20250150773
Methods, systems, and media for determining sound field rotations are provided. In some embodiments, a method for determining sound field rotations involves determining an activity situation of a user. The method may involve determining a user head orientation using at least one sensor of the one or more sensors. The method may involve determining a direction of interest based on the activity situation and the user head orientation. The method may involve determining a rotation of a…
VOLUMETRIC IMMERSIVE EXPERIENCE WITH MULTIPLE VIEWS
Granted: May 8, 2025
Application Number:
20250148699
A multi-view input image covering multiple sampled views is received. A multi-view layered image stack is generated from the multi-view input image. A target view of a viewer to an image space depicted by the multi-view input image is determined based on user pose data. The target view is used to select user pose selected sampled views from among the multiple sampled views. Layered images for the user pose selected sampled views, along with alpha maps and beta scale maps for the user…
SYSTEMS AND METHODS FOR IMPROVED ENTROPY CODING EFFICIENCY
Granted: May 1, 2025
Application Number:
20250142074
Systems and methods for an entropy coding system are described. The entropy coding systems include an encoding apparatus and a decoding apparatus. The encoding apparatus is configured to receive an original input stream comprising a plurality of symbols having a known entropy characteristic according to a probability distribution of each of the symbols appearing in the original input stream, determine an input and respective state for each symbol read from the original input stream,…
METHODS AND DEVICES FOR RENDERING AN AMBISONICS AUDIO SIGNAL
Granted: May 1, 2025
Application Number:
20250142276
The present document describes a method (400) for rendering an ambisonics signal using a loudspeaker arrangement comprising S loudspeakers. The method (400) comprises converting (401) a set of N ambisonics channel signals (111) into a set of unfiltered pre-rendered signals (211), with N>1 and S>1. Furthermore, the method (400) comprises performing (402) near field compensation, referred to as NFC, filtering of M unfiltered pre-rendered signals (211) of the set of unfiltered…
EFFICIENT ORIENTATION TRACKING WITH FUTURE ORIENTATION PREDICTION
Granted: May 1, 2025
Application Number:
20250142282
The present disclosure relates to a method and system for predicting a future orientation of an orientation tracker (100). The method comprising obtaining a sequence of angular velocity samples, each angular velocity sample indicating an angular velocity at a point in time and obtaining a sequence of angular acceleration samples, each angular acceleration sample indicating an acceleration or deceleration of the angular velocity at each point in time. Wherein said method further comprises…
METHOD, APPARATUS OR SYSTEMS FOR PROCESSING AUDIO OBJECTS
Granted: May 1, 2025
Application Number:
20250142285
Diffuse or spatially large audio objects may be identified for special processing. A decorrelation process may be performed on audio signals corresponding to the large audio objects to produce decorrelated large audio object audio signals. These decorrelated large audio object audio signals may be associated with object locations, which may be stationary or time-varying locations. For example, the decorrelated large audio object audio signals may be rendered to virtual or actual speaker…
MANAGEMENT OF PROFESSIONALLY GENERATED AND USER-GENERATED AUDIO CONTENT
Granted: April 24, 2025
Application Number:
20250130756
A system for managing user-generated content (UGC) and professionally generated content (PGC) is disclosed. The system is programmed to receive digital audio data having two channels from a social media platform. The system is programmed to extract spatial features that capture differences in the two channels from the digital audio data. The system is programmed to also extract temporal features, spectral features, and background features from the digital audio data. The system is…
REVERB AND NOISE ROBUST VOICE ACTIVITY DETECTION BASED ON MODULATION DOMAIN ATTENTION
Granted: April 24, 2025
Application Number:
20250131941
A system for detecting speech from reverberant signals is disclosed. The system is programmed to receive spectral temporal amplitude data in the modulation frequency domain. The system is programmed to then enhance the spectral temporal amplitude data by reducing reverberation and other noise as well as smoothing based on certain properties of the spectral temporal spectrogram associated with the spectral temporal amplitude data. Next, the system is programmed to compute various features…
SUPPORTING MULTIPLE TARGET DISPLAY TYPES
Granted: April 24, 2025
Application Number:
20250133224
A sequence of base layer images for a base reference image display and a set of one or more sequences of beta scale maps corresponding to one or more non-base reference image displays are generated. A subset of one or more sequences of beta scale maps is determined in the set of one or more sequences of beta scale maps based at least in part on display capabilities of a target image display. The sequence of base layer images, along with the subset of one or more sequences of beta scale…
TENSOR-PRODUCT B-SPLINE PREDICTION FOR HDR VIDEO IN MOBILE APPLICATIONS
Granted: April 24, 2025
Application Number:
20250133241
Tensor-Product B-splines (TPB) have been shown to improve video quality when used to represent reshaping functions to map reshaped standard dynamic range content into high dynamic range (HDR) content; however, TPB prediction is computationally intensive and may not be supported by legacy devices. Methods and systems for backwards-compatible signaling of TPB-related metadata and a fast TPB prediction method are presented to overcome both of these limitations. Computation overhead for a…
GENERATING CHANNEL AND OBJECT-BASED AUDIO FROM CHANNEL-BASED AUDIO
Granted: April 17, 2025
Application Number:
20250126428
A method of audio processing includes generating a detection score based on the partial loudnesses of a reference audio signal, extracted audio objects, extracted bed channels, a rendered audio signal and a channel-based audio signal. The detection score is indicative of an audio artifact in one or more of the audio objects and the bed channels. The extracted audio objects and extracted bed channels may be modified, in accordance with the detection score, to reduce the audio artifact.
LOUDNESS ADJUSTMENT FOR DOWNMIXED AUDIO CONTENT
Granted: April 17, 2025
Application Number:
20250126427
Audio content coded for a reference speaker configuration is downmixed to downmix audio content coded for a specific speaker configuration. One or more gain adjustments are performed on individual portions of the downmix audio content coded for the specific speaker configuration. Loudness measurements are then performed on the individual portions of the downmix audio content. An audio signal that comprises the audio content coded for the reference speaker configuration and downmix…
METADATA-AIDED REMOVAL OF FILM GRAIN
Granted: April 17, 2025
Application Number:
20250126302
A metadata-aided film-grain removal method and corresponding apparatus. An example embodiment enables a video decoder to substantially fully remove the film grain from a digital video signal that has undergone lossy video compression and then video decompression. Different embodiments may rely only on spatial-domain grain-removal processing, only on temporal-domain grain-removal processing, or on a combination of spatial-domain and temporal-domain grain-removal processing. Both…
METHODS AND SYSTEMS FOR RENDERING OBJECT BASED AUDIO
Granted: April 17, 2025
Application Number:
20250124933
Methods for generating an object based audio program, renderable in a personalizable manner, and including a bed of speaker channels renderable in the absence of selection of other program content (e.g., to provide a default full range audio experience). Other embodiments include steps of delivering, decoding, and/or rendering such a program. Rendering of content of the bed, or of a selected mix of other content of the program, may provide an immersive experience. The program may include…
REPRESENTING SPATIAL AUDIO BY MEANS OF AN AUDIO SIGNAL AND ASSOCIATED METADATA
Granted: April 10, 2025
Application Number:
20250119698
There is provided encoding and decoding methods for representing spatial audio that is a combination of directional sound and diffuse sound. An exemplary encoding method includes inter alia creating a single- or multi-channel downmix audio signal by downmixing input audio signals from a plurality of microphones in an audio capture unit capturing the spatial audio; determining first metadata parameters associated with the downmix audio signal, wherein the first metadata parameters are…
LOCAL AMBIENT COMPENSATION SYSTEMS AND METHODS
Granted: April 10, 2025
Application Number:
20250118276
Novel methods and systems for locally adapting (modifying) an image to compensate for ambient conditions is realized. An adjusted cone response determining based on a minimum target cone response and a delta cone response. A target luminance is then calculated from a local adaptation pooling and the adjusted cone response. Then the image is modified by the target luminance to produce an adapted image.
QUANTUM DOTS AND PHOTOLUMINESCENT COLOR FILTER
Granted: April 10, 2025
Application Number:
20250118236
Image data is received for rendering an image on an image display to a viewer (402). The image data specifies a pixel value of the image for a pixel of the image display to render. The pixel value for the pixel includes multiple component pixel values corresponding to multiple color components of a color space. A color gamut locational value of the pixel value is computed based on two or more component pixel values in the multiple component pixel values of the pixel value specified for…
GENERATING HDR IMAGE FROM CORRESPONDING CAMERA RAW AND SDR IMAGES
Granted: April 10, 2025
Application Number:
20250117904
Guided filtering is applied, with a camera raw image as a guidance image, to a first image to generate an intermediate image. A dynamic range mapping is performed on the intermediate image to generate a second image of a different dynamic range. The second image is used to generate specific local reshaping function index values for selecting specific local reshaping functions. The specific local reshaping functions are applied to the second image to generate a locally reshaped image.
BYGENERATION OF HYBRID IMAGES FOR USE IN CAPTURING PERSONALIZED PLAYBACK-SIDE CONTEXT INFORMATION OF A USER
Granted: April 10, 2025
Application Number:
20250117885
A method may include generating a hybrid image associated with a first interpretation corresponding to a first value of a media parameter and a second interpretation corresponding to a second value of the media parameter. The hybrid image may include a first visibility ratio between the first interpretation and the second interpretation. The method may include refining the hybrid image to create a refined hybrid image that includes a second visibility ratio different than the first…
DEPTH DIFFERENCES IN PLACE OF MOTION VECTORS
Granted: April 3, 2025
Application Number:
20250113057
Input texture images are received. Depth maps are determined for the input texture images. Each depth map in the depth maps includes depth values of pixels represented in a corresponding input texture image in the input texture images. Depth difference maps are generated from the depth maps. The depth difference maps include at least one depth difference map that is generated from two successive depth maps in the depth maps. A video signal encoded with a compressed version of the input…