An example method of video decoding includes receiving a video bitstream comprising a plurality of blocks corresponding to a plurality of pictures, the plurality of blocks including a first block. The method also includes maintaining an ordered transform list for the plurality of blocks, the ordered transform list corresponding to most probable transform options and comprising at least one of: transform type information and transform set information. The method further includes determining, for the first block, an index value for the ordered transform list; and decoding the first block by applying a transform to the first block using information from the transform list at the index value.
H04N 19/12 - Selection from among a plurality of transforms or standards, e.g. selection between discrete cosine transform [DCT] and sub-band transform or selection between H.263 and H.264
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/46 - Embedding additional information in the video signal during the compression process
2.
EFFICIENT CODING OF ATTRIBUTE CONNECTIVITY OF POLYGON MESHES
A bitstream that includes coded information of a mesh with position vertices and non-position vertices is received. The coded information indicates whether a cut is applied to connections of the position vertices of the mesh. Whether the cut is applied is based on a comparison between a total number of the position vertices of the mesh and a total number of the non-position vertices of the mesh. The position vertices of the mesh are processed when the coded information indicates that the cut is applied to the connections of the position vertices of the mesh. The mesh is reconstructed based on whether the position vertices are processed.
This disclosure relates generally to coding and decoding of 3-dimensional (3D) mesh and specifically to merging of boundary vertexes in a symmetric mesh. The disclosure particularly provides methods, systems, and devices for determining whether true boundary vertices are present near a symmetry cut plane during symmetric coding of the 3D mesh, for determining an adaptive distance threshold for merging of boundary vertices, and for determining a list of true boundary vertices in order to reduce incorrect merging of these vertices, thereby reducing cracking artifact near the symmetry plane while preserving true boundaries in the reconstructed mesh
G06T 17/20 - Wire-frame description, e.g. polygonalisation or tessellation
H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
4.
COMPUTING ATTRIBUTE PSNR BASED ON GEOMETRIC ERRORS FOR MESH QUALITY EVALUATION
A first point cloud is determined based on a plurality of first vertices of a reference mesh and a second point cloud is determined based on a plurality of second vertices of a distorted mesh. The distorted mesh is associated with the reference mesh. At least one of a symmetric geometric error or a symmetric attribute error is determined based on a plurality of distances between points of the first point cloud and points of the second point cloud. Each of the plurality of distances is determined between a respective point of the first point cloud and a point of the second point cloud that corresponds to the respective point of the first point cloud. A peak signal to noise ratio (PSNR) is determined based on one of the symmetric geometric error and the symmetric attribute error.
H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
G06T 17/20 - Wire-frame description, e.g. polygonalisation or tessellation
An example method of video coding includes receiving a video bitstream comprising a current block and identifying a first prediction mode for the current block. The method also includes mapping the first prediction mode to a corresponding intra prediction mode, and selecting a first transform set from a plurality of transform sets based on the corresponding intra prediction mode. The method further includes reconstructing the current block using the first transform set.
H04N 19/12 - Selection from among a plurality of transforms or standards, e.g. selection between discrete cosine transform [DCT] and sub-band transform or selection between H.263 and H.264
H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/61 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
6.
MULTI-LAYER CODING FOR HYBRID MACHINE-HUMAN CONSUMPTION
Some aspects of the disclosure provide a method of video decoding. For example, a coded video bitstream is received. The coded video bitstream includes coded information of a video using multi-layer coding. The coded information includes first coded information corresponding to first coded pictures of a first layer serving for a machine task, and second coded information corresponding to second coded pictures of a second layer serving for a human consumption. A consumption type is determined from at least the machine task and the human consumption. Based on the consumption type, at least a reconstructed picture is reconstructed according to at least one of the first coded information and/or the second coded information.
H04N 19/172 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
H04N 19/167 - Position within a video image, e.g. region of interest [ROI]
H04N 19/31 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability in the temporal domain
H04N 19/33 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability in the spatial domain
A video bitstream is received. The video bitstream includes coded information of a current block in a current picture and of a plurality of reference pictures of the current picture in a reference list. The current block includes a plurality of subblocks. A reference block of the current block is determined. The reference block includes a plurality of reference subblocks corresponding to the plurality of subblocks of the current block. A subblock-level motion vector (MV) is determined for a subblock of the plurality of subblocks of the current block. The subblock-level MV is associated with a sum of a plurality of intermediate vectors. Each of the plurality of intermediate vectors is defined between two different respective pictures of the plurality of reference pictures. The subblock of the plurality of subblocks of the current block is reconstructed based on the subblock-level MV.
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/139 - Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/196 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding being specially adapted for the computation of encoding parameters, e.g. by averaging previously computed encoding parameters
H04N 19/52 - Processing of motion vectors by encoding by predictive encoding
Some aspects of the disclosure provide a method of video decoding. For example, a coded video bitstream is received. The coded video bitstream includes at least coded information of a current picture, the current picture includes at least a first block that satisfies an overlapping requirement with a region of interest (ROI), the overlapping requirement requires an overlapping to the ROI to be less than a threshold. The ROI is determined based on the coded information of the current picture. The first block is determined to satisfy the overlapping requirement based on the ROI. At least a first block level syntax of the first block is obtained according a default value that is defined for blocks that satisfies the overlapping requirement. The first block is reconstructed based on the default value of the first block level syntax.
H04N 19/167 - Position within a video image, e.g. region of interest [ROI]
H04N 19/14 - Coding unit complexity, e.g. amount of activity or edge presence estimation
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
9.
LIST CONSTRUCTION AND SIGNALING OF TRANSFORM TYPES AND TRANSFORM SETS
An example method of video decoding includes receiving a video bitstream comprising a plurality of blocks corresponding to a plurality of pictures, the plurality of blocks including a first block. The method also includes maintaining an ordered transform list for the plurality of blocks, the ordered transform list corresponding to most probable transform options and comprising at least one of: transform type information and transform set information. The method further includes determining, for the first block, an index value for the ordered transform list; and decoding the first block by applying a transform to the first block using information from the transform list at the index value.
H04N 19/10 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
H04N 19/11 - Selection of coding mode or of prediction mode among a plurality of spatial predictive coding modes
H04N 19/12 - Selection from among a plurality of transforms or standards, e.g. selection between discrete cosine transform [DCT] and sub-band transform or selection between H.263 and H.264
H04N 19/46 - Embedding additional information in the video signal during the compression process
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/50 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
H04N 19/60 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
An example method of video coding includes receiving a video bitstream comprising a current block and a syntax element and identifying an intra prediction mode for the current block. The method also includes parsing the syntax element to determine whether to change the intra prediction mode when deriving a transform set index, and deriving the transform set index for the current block based on the intra prediction mode and the parsed syntax element. The method further includes selecting a first transform set from a plurality of transform sets according to the transform set index, and reconstructing the current block using the first transform set.
H04N 19/61 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
H04N 19/18 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a set of transform coefficients
H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
H04N 19/11 - Selection of coding mode or of prediction mode among a plurality of spatial predictive coding modes
A method of video decoding includes receiving a coded video bitstream comprising coded information of a coded picture and receiving a supplemental enhancement information (SEI) message associated with the coded picture. The SEI message indicates one or more of (i) a type of film grain, (ii) a purpose of the film grain, and (iii) essentiality of the film grain that is applied to a first region including one or more first samples in the coded picture. The method of video decoding includes reconstructing the coded picture that is associated with the SEI message.
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
H04N 19/85 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
H04N 19/167 - Position within a video image, e.g. region of interest [ROI]
H04N 19/172 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
H04N 19/169 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
H04N 19/174 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a slice, e.g. a line of blocks or a group of blocks
H04N 19/117 - Filters, e.g. for pre-processing or post-processing
A method of video decoding includes reconstructing a coded picture including at least a first sample. The coded picture is associated with a Supplemental Enhancement Information (SEI) message for a film grain synthesis process to be applied to the reconstructed picture. The SEI message for the film grain synthesis process includes a syntax element indicating that alpha channel information is used in the film grain synthesis process. The method of video decoding includes applying the film grain synthesis process to the first sample based at least on the alpha channel information.
H04N 19/46 - Embedding additional information in the video signal during the compression process
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
H04N 21/00 - Selective content distribution, e.g. interactive television or video on demand [VOD]
H04N 7/12 - Systems in which the television signal is transmitted via one channel or a plurality of parallel channels, the bandwidth of each channel being less than the bandwidth of the television signal
Some aspects of the disclosure provide a method of video decoding. For example, a coded video bitstream is received. The coded video bitstream includes coded information of a video for a machine task. Among coding tools of a video codec, a first subset of the coding tools is selected. The first subset of the coding tools is used for coding the video for the machine task. Decoded information is generated from the coded video bitstream based on the first subset of the coding tools. The machine task is performed using the decoded information.
H04N 19/103 - Selection of coding mode or of prediction mode
H04N 19/136 - Incoming video signal characteristics or properties
H04N 19/157 - Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
H04N 19/184 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being bits, e.g. of the compressed video stream
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
Some aspects of the disclosure provide a method of video decoding. In some examples, a coded video bitstream is received. The coded video bitstream includes at least coded information of a block in a picture of a video. Transform coefficients are extracted from the coded information of the block. For example, the coded information of the block is parsed and decoded to extract the transform coefficients. An inverse transform is applied on the transform coefficients to generate scaled residual values. An inverse scaling operation is applied on the scaled residual values based on one or more scaling values to generate restored residual values. The block is reconstructed according to the restored residual values. In an example, the restored residual values are combined with prediction of samples of the block to generate the reconstructed samples of the block.
H04N 19/59 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial sub-sampling or interpolation, e.g. alteration of picture size or resolution
H04N 19/139 - Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/18 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a set of transform coefficients
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
15.
METHOD AND APPARATUS FOR TEMPORAL INTERPOLATED PREDICTION MODE IMPROVEMENT
A video decoder is provided for the decoding of a video bitstream encoded in a temporal interpolated prediction (TIP) mode. First and second motion vectors pointing to respective reference frames, or reference pictures within those frames, are generated for a block of a current picture. The motion vectors are then refined by application of a decoder-side motion vector refinement (DMVR) process, based on a bilateral matching process, and the refined motion vectors are used to decode the block. The refinement may more specifically involve consideration of candidates for the refined motion vectors, selected by the bilateral matching. The refinement may be applied at both block and sub-block divisions of the current picture.
H04N 19/44 - Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
H04N 19/137 - Motion inside a coding unit, e.g. average field, frame or block difference
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
16.
SEI MESSAGE FOR CARRIAGE OF TEXT DATA FOR GENERATIVE ARTIFICIAL INTELLIGENCE APPLICATIONS IN VIDEO STREAMS
A method of processing a video bitstream includes receiving the video bitstream comprising (i) one of a picture and a video and (ii) a supplemental enhancement information (SEI) message associated with the one of the picture and the video. The SEI message includes text data purposed for use with a generative artificial intelligence (Al) process. The method of processing the video bitstream includes extracting the text data from the SEI message. The SEI message does not indicate whether the one of the picture and the video has been modified by another generative Al process.
H04N 19/46 - Embedding additional information in the video signal during the compression process
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
H04N 19/36 - Scalability techniques involving formatting the layers as a function of picture distortion after decoding, e.g. signal-to-noise [SNR] scalability
H04N 21/00 - Selective content distribution, e.g. interactive television or video on demand [VOD]
H04N 7/12 - Systems in which the television signal is transmitted via one channel or a plurality of parallel channels, the bandwidth of each channel being less than the bandwidth of the television signal
H04N 21/80 - Generation or processing of content or additional data by content creator independently of the distribution processContent per se
17.
BIDIRECTIONAL COMMUNICATION WITH INFRASTRUCTURE AS CODE SYSTEM
A method is provided, in which infrastructure configuration parameters received from a configuration management database are converted into an infrastructure-as-code configuration language file. Infrastructure configuration modifications are applied to the infrastructure-as-code configuration language file. The modified infrastructure-as-code configuration language file is forwarded to an infrastructure-as-a-service provider. Provisional infrastructure-as-code parameters output by the infrastructure-as-a-service provider in response to the modified infrastructure-as-code configuration language file are converted to updated infrastructure configuration parameters. The updated infrastructure configuration parameters are transmitted for storage in the configuration management database.
A method of processing a video bitstream includes receiving the video bitstream comprising (i) one of a picture and a video and (ii) a supplemental enhancement information (SEI) message associated with the one of the picture and the video. The SEI message includes text data purposed for use with a generative artificial intelligence (AI) process. The method of processing the video bitstream includes extracting the text data from the SEI message. The SEI message does not indicate whether the one of the picture and the video has been modified by another generative AI process.
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
G06V 10/82 - Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
H04N 19/117 - Filters, e.g. for pre-processing or post-processing
H04N 19/184 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being bits, e.g. of the compressed video stream
H04N 19/46 - Embedding additional information in the video signal during the compression process
19.
SUPPLEMENTAL ENHANCEMENT INFORMATION (SEI) MESSAGE FOR FILM GRAIN SYNTHESIS EXTENSION
A method of video decoding includes reconstructing a coded picture including at least a first sample. The coded picture is associated with a Supplemental Enhancement Information (SEI) message for a film grain synthesis process to be applied to the reconstructed picture. The SEI message for the film grain synthesis process includes a syntax element indicating that alpha channel information is used in the film grain synthesis process. The method of video decoding includes applying the film grain synthesis process to the first sample based at least on the alpha channel information.
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
H04N 19/167 - Position within a video image, e.g. region of interest [ROI]
H04N 19/172 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
A method of video decoding includes receiving a coded video bitstream comprising coded information of a coded picture and receiving a supplemental enhancement information (SEI) message associated with the coded picture. The SEI message indicates one or more of (i) a type of film grain, (ii) a purpose of the film grain, and (iii) essentiality of the film grain that is applied to a first region including one or more first samples in the coded picture. The method of video decoding includes reconstructing the coded picture that is associated with the SEI message.
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
H04N 19/167 - Position within a video image, e.g. region of interest [ROI]
H04N 19/172 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
An example method of video coding includes receiving a video bitstream comprising a current block and a syntax element and identifying an intra prediction mode for the current block. The method also includes parsing the syntax element to determine whether to change the intra prediction mode when deriving a transform set index, and deriving the transform set index for the current block based on the intra prediction mode and the parsed syntax element. The method further includes selecting a first transform set from a plurality of transform sets according to the transform set index, and reconstructing the current block using the first transform set.
H04N 19/12 - Selection from among a plurality of transforms or standards, e.g. selection between discrete cosine transform [DCT] and sub-band transform or selection between H.263 and H.264
H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/61 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
22.
Joint Signaling Method for Motion Vector Difference
A method for encoding and decoding of motion vector difference for inter-predicting a video block is provided. The method includes receiving a coded video bitstream; extracting, from the coded video bitstream, a flag indicating whether an inter-predication mode is a JOINT_NEWMV mode for a current block in a current frame, the JOINT_NEWMV mode indicating that a first delta motion vector (MV) for a first reference frame from a reference list 0 and a second delta MV for a second reference frame from a reference list 1 are jointly signaled; in response to the flag indicating that the inter-predication mode is the JOINT_NEWMV mode, extracting a joint delta motion vector (MV) for the current block, and deriving the first delta MV and the second delta MV based on the joint delta MV; and decoding the current block based on the first delta MV and the second delta MV.
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/139 - Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
A method for video encoding includes determining an inferable condition for a flag of a specific prediction mode for a current block. The specific prediction mode is one of a plurality of inter picture prediction modes, the inferable condition is based on whether the specific prediction mode and other modes in the plurality of inter prediction modes are allowed, and the other modes are at a lower level than the specific prediction mode in a tree structure. Encoding of the flag in a video bitstream is skipped when the inferable condition indicates that the flag is inferable. The flag is encoded in the video bitstream when the inferable condition indicates uncertainty for inferring the flag. The current block is encoded according to the specific prediction mode when the flag is indicative of an application of the specific prediction mode on the current block.
H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
H04N 19/109 - Selection of coding mode or of prediction mode among a plurality of temporal predictive coding modes
H04N 19/169 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
H04N 19/172 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
H04N 19/174 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a slice, e.g. a line of blocks or a group of blocks
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/184 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being bits, e.g. of the compressed video stream
H04N 19/52 - Processing of motion vectors by encoding by predictive encoding
24.
Chroma from Luma Prediction Using Neighbor Luma Samples
This disclosure relates to video processing that includes a video processing device that: determines that a Chroma from Luma (CfL) prediction mode is to be applied to a luma block in a received coded bitstream; generates a neighbor luma average for the luma block by averaging a set of reconstructed luma samples, wherein the set of reconstructed luma samples comprises a plurality of reconstructed neighbor luma samples in at least one neighbor luma block that neighbors the luma block; generates an alternating current (AC) contribution of a plurality of prediction samples of a chroma block co-located with the luma block based on a plurality of luma samples in the luma block and the neighbor luma average; and reconstructs the chroma block at least by applying the CfL prediction mode based on the AC contribution.
H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/593 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
A method is provided for adjusting a perceived map location of a first player in a video game. The method includes determining whether the perceived map location of a virtual game asset of a first player of the game starting from a first location on a map of a virtual environment is to be displayed as starting from a second location on the map, the first location is symmetrical from the second location about a reference point on the map. The method includes, based on a determination that the perceived map location is to be displayed as starting from the second location, adjusting a location of the virtual game asset on the map to an adjusted location for display on a device displayed map according to the second location. The method further includes displaying the virtual game asset of the first player at the adjusted location on the device displayed map.
A63F 13/5372 - Controlling the output signals based on the game progress involving additional visual information provided to the game scene, e.g. by overlay to simulate a head-up display [HUD] or displaying a laser sight in a shooting game using indicators, e.g. showing the condition of a game character on screen for tagging characters, objects or locations in the game scene, e.g. displaying a circle under the character controlled by the player
A63F 13/5258 - Changing parameters of virtual cameras by dynamically adapting the position of the virtual camera to keep a game object or game character in its viewing frustum, e.g. for tracking a character or a ball
A63F 13/5378 - Controlling the output signals based on the game progress involving additional visual information provided to the game scene, e.g. by overlay to simulate a head-up display [HUD] or displaying a laser sight in a shooting game using indicators, e.g. showing the condition of a game character on screen for displaying an additional top view, e.g. radar screens or maps
26.
BIDIRECTIONAL COMMUNICATION WITH INFRASTRUCTURE AS CODE SYSTEM
A method is provided, in which infrastructure configuration parameters received from a configuration management database are converted into an infrastructure-as-code configuration language file. Infrastructure configuration modifications are applied to the infrastructure-as-code configuration language file. The modified infrastructure-as-code configuration language file is forwarded to an infrastructure-as-a-service provider. Provisional infrastructure-as-code parameters output by the infrastructure-as-a-service provider in response to the modified infrastructure-as-code configuration language file are converted to updated infrastructure configuration parameters. The updated infrastructure configuration parameters are transmitted for storage in the configuration management database.
A method is provided for improving multiplayer user experience in an online video game utilizing a lockstep engine. The method includes receiving user input information from at least one of a plurality of users of a match of the video game that utilizes the lockstep engine. Each of the user input information from the at least one of the plurality of users being associated with a current frame of a plurality of frames of the lockstep engine based on the respective user input information being received within a current time interval of the current frame. The method also includes sending the current frame of the user input information irrespective of whether the user input information is received from each of the plurality of users within the current time interval. Clients no longer need to be in the game at the same time for the game to proceed.
A63F 13/75 - Enforcing rules, e.g. detecting foul play or generating lists of cheating players
A63F 13/34 - Interconnection arrangements between game servers and game devicesInterconnection arrangements between game devicesInterconnection arrangements between game servers using peer-to-peer connections
A63F 13/355 - Performing operations on behalf of clients with restricted processing capabilities, e.g. servers transform changing game scene into an encoded video stream for transmitting to a mobile phone or a thin client
An aspect of the disclosure provides a method of mesh decoding. For example, a bitstream that includes coded information of a mesh frame is received. A syntax element is parsed from the bitstream, the syntax element indicates whether an adaptive linear wavelet transform is used, the adaptive linear wavelet transform applies different weight values to different neighboring vertices of a vertex in a wavelet transform of attribute values associated with vertices of the mesh frame. When the syntax element indicates a use of the adaptive linear wavelet transform, at least a first vertex in the vertices of the mesh frame is reconstructed according to the adaptive linear wavelet transform.
H04N 19/63 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding using sub-band based transform, e.g. wavelets
H04N 19/117 - Filters, e.g. for pre-processing or post-processing
H04N 19/167 - Position within a video image, e.g. region of interest [ROI]
H04N 19/172 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
H04N 19/196 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding being specially adapted for the computation of encoding parameters, e.g. by averaging previously computed encoding parameters
H04N 19/61 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
29.
IMPLICIT MASKED BLENDING MODE IMPROVEMENT WITH HIGH LEVEL SYNTAX
An example method of video encoding includes receiving video data comprising first block. The method also includes determining that the first block is to be encoded using first and second reference blocks and signaling a high-level syntax (HLS) flag in a video bitstream that indicates whether an implicit masked blending mode is enabled for the first block. When the HLS flag having a first value, encoding the first block using a first set of weighting factors for reference values from the first reference block and the second reference block. When (i) the HLS flag having a second value and (ii) at least one of the first reference block and the second reference block is at least partially outside of a corresponding reference boundary, encoding the first block using a second set of weighting factors for the reference values from the first reference block and the second reference block.
H04N 19/44 - Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
Systems and methods for decoding are provided, a method includes: encoding a first sub-picture of a picture, independently from a second sub-picture of the picture, using sub-picture, tile group, and tile partitioning: and sending, to at least one decoder, at least one coded sub-bitstream of a coded video stream that includes the first sub-picture and the second sub-picture, wherein (i) the first sub-picture is a first rectangular region of the picture and the second sub-picture is a second rectangular region of the picture, the second rectangular region being different from the first rectangular region, (ii) the first sub-picture includes a first plurality of tiles, and the second sub-picture includes at least one tile, (iii) the first sub-picture and the second sub-picture do not share a common tile, and (iv) the first plurality of tiles of the first sub-picture are grouped into at least two first tile groups.
H04N 19/119 - Adaptive subdivision aspects e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
H04N 19/117 - Filters, e.g. for pre-processing or post-processing
H04N 19/184 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being bits, e.g. of the compressed video stream
A method is provided for replaying a match in a video game. The method includes receiving player input information of the match in the video game that utilizes a lockstep game engine. The player input information includes player input data for each of a plurality of frames of the lockstep game engine. The method also includes generating a plurality of game states of the match in the video game based on the player input data for each of the plurality of frames. The method further includes storing the plurality of game states of the match in at least one memory. The method also includes receiving a selection of a playback point of the match. The method further includes generating the plurality of video frames of the match using the stored plurality of game states based on the selected playback point.
Some aspects of the disclosure provide a method of video decoding. In the method, a coded video bitstream comprising coded information of a current block in a current picture is received. The coded information of the current block indicates an intra prediction mode for the current block. One or more in-loop filters is applied on one or more reconstructed samples in the current picture to generate one or more filtered reconstructed samples. The current block is reconstructed using the intra prediction mode and based on the one or more filtered reconstructed samples.
H04N 19/82 - Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation involving filtering within a prediction loop
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/55 - Motion estimation with spatial constraints, e.g. at image or region borders
33.
SYSTEMS AND METHODS FOR DECODER-SIDE MOTION VECTOR REFINEMENT SUBSAMPLING
An example method of video coding includes receiving a video bitstream comprising a plurality of blocks, including a current block. The method also includes identifying a current subblock in the current block and deriving a subsampled subblock by sub-sampling the current subblock. The method further includes deriving, using the subsampled subblock, a set of subblock motion vectors for the current subblock, and deriving a set of refined subblock motion vectors from the set of subblock motion vectors. Additionally, the method includes reconstructing the current subblock using the derived set of refined subblock motion vectors.
H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
H04N 19/137 - Motion inside a coding unit, e.g. average field, frame or block difference
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/139 - Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
Some aspects of the disclosure provide a method of mesh processing. The method includes receiving a bitstream including coded information of a polygon mesh, the polygon mesh includes vertices that are connected into polygon faces, the coded information includes dual- degree connectivity information of the polygon mesh and residuals of geometry predictors for the polygon mesh. The method also includes reconstructing at least a first connectivity of a first polygon face according to the dual-degree connectivity information, the first connectivity of the first polygon face indicates connections of first vertices in the vertices into the first polygon face. Further, the method includes triangulating the first polygon face into a first set of triangles and determining first geometry information of the first vertices based on the first set of triangles.
An example method of video coding includes receiving a video bitstream comprising a plurality of blocks, including a current block. The method also includes identifying a current subblock in the current block and deriving a subsampled subblock by sub-sampling the current subblock. The method further includes deriving, using the subsampled subblock, a set of subblock motion vectors for the current subblock, and deriving a set of refined subblock motion vectors from the set of subblock motion vectors. Additionally, the method includes reconstructing the current subblock using the derived set of refined subblock motion vectors.
H04N 19/137 - Motion inside a coding unit, e.g. average field, frame or block difference
H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
36.
COMBINE DUAL-DEGREE BASED CONNECTIVITY CODING AND TRIANGLE BASED GEOMETRY CODING FOR POLYGON MESH COMPRESSION
Some aspects of the disclosure provide a method of mesh processing. The method includes receiving a bitstream including coded information of a polygon mesh, the polygon mesh includes vertices that are connected into polygon faces, the coded information includes dual-degree connectivity information of the polygon mesh and residuals of geometry predictors for the polygon mesh. The method also includes reconstructing at least a first connectivity of a first polygon face according to the dual-degree connectivity information, the first connectivity of the first polygon face indicates connections of first vertices in the vertices into the first polygon face. Further, the method includes triangulating the first polygon face into a first set of triangles and determining first geometry information of the first vertices based on the first set of triangles.
H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
G06T 17/20 - Wire-frame description, e.g. polygonalisation or tessellation
Neighboring blocks of a current block include first blocks adjacent to one of a top side, an above left corner, and an above right corner of the current block and second blocks adjacent to one of a left side and a below left corner of the current block. Whether one or more of the first blocks and the current block are in a same coding tree unit (CTU) is determined. Based on the one or more of the first blocks and the current block being in the same CTU, a respective intra mode associated with each of the one or more of the first blocks is added into a most probable mode (MPM) list for the current block based on a sequence of conditions. A respective intra mode associated with each of the second blocks is added into the MPM list based on the sequence of conditions.
H04N 19/196 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding being specially adapted for the computation of encoding parameters, e.g. by averaging previously computed encoding parameters
H04N 19/107 - Selection of coding mode or of prediction mode between spatial and temporal predictive coding, e.g. picture refresh
H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
An example method of video coding includes receiving a video bitstream comprising a plurality of blocks. The method also includes deriving a set of subblock motion vectors for a current subblock in a first block of the plurality of blocks, and deriving a set of refined subblock motion vectors for the current subblock by applying an optical flow refinement that uses one or more values that are outside of the current subblock, where a block size for the optical flow refinement is a block size of the current subblock. The method further includes reconstructing the current subblock using the set of refined subblock motion vectors.
H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
H04N 19/137 - Motion inside a coding unit, e.g. average field, frame or block difference
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/139 - Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
Methods and apparatuses for performing chroma from luma (CfL) intra prediction, including: receiving a current block from a coded video bitstream; obtaining, from the coded video bitstream, a syntax element indicating which of two or more downsampling filters is used for predicting the current block in a CfL intra prediction mode; in response to the syntax element indicating that a first downsampling filter is used for the current block: determining a plurality of filter coefficients according to the first downsampling filter; and downsampling the current block based on the determined plurality of coefficients using a first number of sampling positions; in response to the syntax element indicating that a second downsampling filter is used for the current block: determining the plurality of filter coefficients according to the second downsampling filter; downsampling the current block based on the determined plurality of coefficients using a second number of sampling positions.
H04N 19/593 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
H04N 19/117 - Filters, e.g. for pre-processing or post-processing
H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
40.
SYSTEMS AND METHODS FOR OPTICAL FLOW-BASED MOTION VECTOR REFINEMENT
An example method of video coding includes receiving a video bitstream comprising a plurality of blocks. The method also includes deriving a set of subblock motion vectors for a current subblock in a first block of the plurality of blocks, and deriving a set of refined subblock motion vectors for the current subblock by applying an optical flow refinement that uses one or more values that are outside of the current subblock, where a block size for the optical flow refinement is a block size of the current subblock. The method further includes reconstructing the current subblock using the set of refined subblock motion vectors.
H04N 19/117 - Filters, e.g. for pre-processing or post-processing
H04N 19/137 - Motion inside a coding unit, e.g. average field, frame or block difference
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/80 - Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation
A method of encoding includes receiving a polygon mesh that includes a plurality of vertices; subdividing the polygon mesh to generate a plurality sub-vertices; determining a first normal vector of a first vertex in the polygon mesh and a second normal vector of a second vertex in the polygon mesh, the first vertex and the second vertex defining an edge in the polygon mesh; moving a sub-vertex from the plurality of sub-vertices located on the edge using the first normal vector and the second normal vector; determining a displacement between the moved sub-vertex and a reference vertex from the plurality of vertices; and generating a bitstream including the displacement.
A bitstream including coded information of a current block and neighboring blocks of the current block in a current picture is received. Whether a merge mode is enabled for the current block is determined based on whether one or more of the neighboring blocks are coded by an intra template matching mode. When the merge mode is enabled for the current block, a merge candidate list that includes the one or more of the neighboring blocks of the current block is generated. A reference block vector is determined based on template matching information of a merge candidate selected from the merge candidate list. The reference block vector is refined to obtain a refined reference block vector for the current block. The current block is reconstructed based on a reference block indicated by the refined reference block vector.
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/20 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video object coding
H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
H04N 19/44 - Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
H04N 19/46 - Embedding additional information in the video signal during the compression process
43.
MERGE AND BLOCK VECTOR REFINEMENT FOR INTRA TEMPLATE MATCHING PREDICTION
A bitstream including coded information of a current block and neighboring blocks of the current block in a current picture is received. Whether a merge mode is enabled for the current block is determined based on whether one or more of the neighboring blocks are coded by an intra template matching mode. When the merge mode is enabled for the current block, a merge candidate list that includes the one or more of the neighboring blocks of the current block is generated. A reference block vector is determined based on template matching information of a merge candidate selected from the merge candidate list. The reference block vector is refined to obtain a refined reference block vector for the current block. The current block is reconstructed based on a reference block indicated by the refined reference block vector.
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
44.
BLOCK BASED WEIGHTING FACTOR FOR JOINT MOTION VECTOR DIFFERENCE CODING MODE
Aspects of the disclosure include a method for video decoding that includes receiving coding information for a block indicating that the block is coded with a joint motion vector difference (JMVD) coding mode and a compound weighted prediction mode and includes scaling factor information of the JMVD coding mode. When the scaling factor information indicates that each scaling factor of vertical and horizontal components of a plurality of MVDs associated with respective reference frames of the block is 1, the method includes determining a weighting factor of the compound weighted prediction mode based on a weighting factor index signaled in a bitstream and a list of weighting factors. The method includes determining, using the JMVD coding mode, motion information associated with the respective reference frames based on the scaling factors and reconstructing, using the compound weighted prediction mode, the block based on the motion information and the determined weighting factor.
H04N 19/44 - Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/137 - Motion inside a coding unit, e.g. average field, frame or block difference
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
45.
METHOD AND APPARATUS FOR CONSISTENCY DETECTION AND RESOLUTION IN AUTOMATIC DIALOGUE SYSTEMS
A method for performing consistency detection in a dialogue system includes training a model based on an annotated dataset for consistency detection to generate a trained model. The method further includes receiving a plurality of utterances from one or more users. The method includes determining, using the trained model, whether the plurality of utterances include an inconsistency. The method includes, based on the determination that the plurality of utterances include the inconsistency, generating, using the trained model, one or more clarifying responses.
A method for video coding includes obtaining a coding block of a video bitstream, determining whether a joint coding of motion vector difference (JMVD) is used for predicting the coding block, obtaining, based on determining that the JMVD is used selected for predicting the coding block, scaling factors, deriving a motion vector difference (MVD) for one or more reference frame lists based on an application of the scaling factors to one or more components of the JMVD along one or more pre-defined directions, and reconstructing the coding block based on at least the derived MVD.
H04N 19/137 - Motion inside a coding unit, e.g. average field, frame or block difference
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/46 - Embedding additional information in the video signal during the compression process
A method performed by a video decoder includes receiving a coded video bitstream including a current picture that includes at least one block. The method includes determining that the at least one block is to be predicted in a reconstruction-reordered intra block copy (RR-IBC) mode. The method includes obtaining a syntax element from the at least one block, the syntax element indicating a flip mode. The method includes determining whether reconstruction flip is applied to the at least one block. The method includes, in response to determining the reconstruction flip is applied to the at least one block, predicting a flip pattern for the at least one block based on neighboring reconstruction samples of the at least block and a corresponding reference block of the at least one block. The method further includes decoding the at least one block based on the flip mode and the predicted flip pattern.
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
H04N 19/154 - Measured or subjectively estimated visual quality after decoding, e.g. measurement of distortion
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
48.
INSTANCE-LEVEL ADAPTIVE PROPULSION OF EXTERNAL KNOWLEDGE (IAPEK)
There is included a method and apparatus comprising computer code for instance-wise adaptive knowledge injection in a pre-trained language model (PTLM) including determining a necessity of external knowledge in a plurality of queries of a first dataset based on a likelihood that a respective query is solved by internal knowledge of a target model. Then, the one or more queries determined to need external knowledge may be augmented with pieces of external knowledge. A combined dataset may be generated by combining the first dataset and the one or more augmented queries, and the combined dataset may be applied to the target model.
In a method, a similarity value is calculated based on samples in at least one of a plurality of candidate templates of a current block and samples of the current block corresponding to the samples in the at least one of the plurality of candidate templates. The plurality of candidate templates includes at least one of a left template, a top template, and an L-shaped template that is adjacent to the current block. A template of the current block is determined as one of the plurality of candidate templates based on whether the similarity value is larger than a threshold value or corresponds to the one of the plurality of candidate templates. A prediction model is determined based on the template of the current block and a template of a reference block of the current block. The current block is encoded based on the prediction model and the reference block.
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/11 - Selection of coding mode or of prediction mode among a plurality of spatial predictive coding modes
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/593 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
A bitstream is received. The bitstream includes coded information of a current block and a template of the current block in a current picture. The template including neighboring samples of the current block. At least one of a size of the template and an offset of the template is determined according to an angle indicated by an initial intra prediction mode of the current block. A refined intra prediction mode for the current block is determined based on a template prediction of the template and a template reconstruction of the template. The current block is reconstructed based on the refined intra prediction mode.
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/147 - Data rate or code amount at the encoder output according to rate distortion criteria
H04N 19/51 - Motion estimation or motion compensation
A method and apparatus comprising computer code configured to cause a processor or processors to obtain video data including at least one coded picture; identify, by a decoder, at least one first supplementary information enhancement (SEI) message, the at least one first SEI message indicating copyright information of the at least one coded picture; and decoding, by the decoder, the video data based on the first SEI message.
H04N 21/858 - Linking data to content, e.g. by linking an URL to a video object or by creating a hotspot
H04N 19/184 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being bits, e.g. of the compressed video stream
A method and apparatus comprising computer code for video processing, the method including setting a first value of a text description purpose parameter in a bitstream, the first value of the text description purpose parameter indicating a type of information included in a text description information string in the bitstream; setting artificial intelligence (AI) marking information associated with one or more pictures in the bitstream as the text description information string when the first value indicates that the type of information included in the text description information string comprises marking information associated with one or more artificial intelligence processes used; signaling the text description purpose parameter in the bitstream; and signaling the text description information string.
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
H04N 19/172 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
H04N 19/189 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding
A bitstream is received. The bitstream includes coded information of a current block and a template of the current block in a current picture. The template including neighboring samples of the current block. At least one of a size of the template and an offset of the template is determined according to an angle indicated by an initial intra prediction mode of the current block. A refined intra prediction mode for the current block is determined based on a template prediction of the template and a template reconstruction of the template. The current block is reconstructed based on the refined intra prediction mode.
H04N 19/11 - Selection of coding mode or of prediction mode among a plurality of spatial predictive coding modes
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
A method and apparatus comprising computer code configured to cause a processor or processors to obtain video data including at least one coded picture; identify, by a decoder, at least one first supplementary information enhancement (SEI) message, the at least one first SEI message indicating copyright information of the at least one coded picture; and decoding, by the decoder, the video data based on the first SEI message.
H04N 19/44 - Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
55.
SYSTEMS AND METHODS FOR IMPLICIT DERIVATION IN A RECURSIVE INTRA REGION
An example method of video coding includes receiving a video bitstream including a plurality of coding blocks; identifying, based on a first indicator in the video bitstream, a coding region that comprises two or more blocks of the plurality of coding blocks, where at least one block in the coding region is encoded in an intra prediction mode and at least one block in the coding region is encoded in a non-intra prediction mode. For a last block in the coding region, when all preceding blocks in the coding region are encoded in the intra prediction mode, the method includes inferring a prediction mode of the last block as being an inter prediction mode without receiving an inter prediction mode indicator for the last block; and reconstructing the last block using the inferred prediction mode.
An example method of video coding includes receiving a video bitstream including a plurality of blocks; receiving a video bitstream that includes a current picture, the current picture including a plurality of blocks that includes a current block. The method includes constructing an intra prediction mode table by adding an entry corresponding to an intra prediction mode to the intra prediction mode table based on one or more intra prediction modes used by one or more previously decoded blocks; and deriving an intra mode prediction list for the current block by selecting one or more entries from the constructed intra prediction mode table. The method also includes reconstructing the current block using an intra prediction mode from the intra prediction mode list.
H04N 1/417 - Systems or arrangements allowing the picture to be reproduced without loss or modification of picture-information using predictive or differential encoding
H04N 19/00 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
H04N 19/10 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/11 - Selection of coding mode or of prediction mode among a plurality of spatial predictive coding modes
57.
SYSTEMS AND METHODS FOR APPLYING NON-SEPARABLE TRANSFORMS ON INTER PREDICTION RESIDUALS
The various implementations described herein include methods and systems for coding video. In one aspect, a method includes receiving a video bitstream that includes a set of inter mode encoded blocks and a corresponding set of transform coefficients. The method includes deriving a set of inter mode residual blocks from the set of transform coefficients. The method includes determining, according to a value of a first indicator in the video bitstream, whether one or more non-separable transform kernels are to be applied to the set of inter mode residual blocks. The method includes applying a first non-separable transform kernel when the indicator has a first value, and forgoing applying the first non-separable transform kernels when the indicator has a second value. The method also includes reconstructing a set of video blocks using the set of inter mode residual blocks and a corresponding set of prediction blocks.
H04N 19/103 - Selection of coding mode or of prediction mode
H04N 19/11 - Selection of coding mode or of prediction mode among a plurality of spatial predictive coding modes
H04N 19/12 - Selection from among a plurality of transforms or standards, e.g. selection between discrete cosine transform [DCT] and sub-band transform or selection between H.263 and H.264
H04N 19/13 - Adaptive entropy coding, e.g. adaptive variable length coding [AVLC] or context adaptive binary arithmetic coding [CABAC]
H04N 19/157 - Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
H04N 19/18 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a set of transform coefficients
58.
METHOD AND APPARATUS FOR CONSISTENCY DETECTION AND RESOLUTION IN AUTOMATIC DIALOGUE SYSTEMS
A method for performing consistency detection in a dialogue system includes training a model based on an annotated dataset for consistency detection to generate a trained model. The method further includes receiving a plurality of utterances from one or more users. The method includes determining, using the trained model, whether the plurality of utterances include an inconsistency. The method includes, based on the determination that the plurality of utterances include the inconsistency, generating, using the trained model, one or more clarifying responses.
H04L 51/02 - User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail using automatic reactions or user delegation, e.g. automatic replies or chatbot-generated messages
59.
SYSTEMS AND METHODS FOR GENERATING INTRA MODE TABLES
An example method of video coding includes receiving a video bitstream including a plurality of blocks; receiving a video bitstream that includes a current picture, the current picture including a plurality of blocks that includes a current block. The method includes constructing an intra prediction mode table by adding an entry corresponding to an intra prediction mode to the intra prediction mode table based on one or more intra prediction modes used by one or more previously decoded blocks; and deriving an intra mode prediction list for the current block by selecting one or more entries from the constructed intra prediction mode table. The method also includes reconstructing the current block using an intra prediction mode from the intra prediction mode list.
H04N 19/11 - Selection of coding mode or of prediction mode among a plurality of spatial predictive coding modes
H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
Processing circuitry decodes, from a bitstream carrying a video including a current picture, a signal indicative of a subblock based intra block copy (IBC) merge mode for predicting a current block in the current picture. The processing circuitry determines a predictor offset (also referred to as offset BV) associated with the current block. The predictor offset indicates a predictor block in the current picture corresponding to the current block, the predictor block includes a plurality of predictor subblocks respectively corresponding to a plurality of subblocks in the current block. The processing circuitry determines a first block vector of a first subblock according to a first predictor block vector associated with a first predictor subblock. The processing circuitry reconstructs the first subblock in the current block according to the first block vector that points to a reference block for the first subblock in the current picture.
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
61.
METHOD AND SYSTEM OF VIDEO CODING USING A SUBSET OF AVAILABLE INTRA PREDICTION MODES FOR MULTIPLE REFERENCE LINES
An example video decoding method includes obtaining a video bitstream that includes a plurality of encoded blocks including a current block. When a multiple reference line intra prediction (MRLP) mode is enabled for the current block and a non-adjacent reference line is used for the current block, a first allowed intra prediction mode (AIPM) set is determined for the current block. The first AIPM includes a set of one or more intra prediction modes of neighboring blocks of the current block, The first AIPM also includes a set of one or more intra prediction modes derived based on a signaled nominal angular mode and an angle delta value. The method further includes reconstructing the current block using an intra prediction mode from the first AIPM set.
H04N 19/593 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/184 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being bits, e.g. of the compressed video stream
H04N 19/44 - Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
62.
HARMONIZED DESIGN FOR OFFSET BASED REFINEMENT AND MULTIPLE REFERENCE LINE SELECTION
Methods, apparatus, and computer readable storage medium for offset refinement for intra prediction and multiple reference line intra prediction in video decoding. The method includes receiving, by a device, a coded video bitstream for a block. The device includes a memory storing instructions and a processor in communication with the memory. The method further includes determining, by the device, whether offset refinement for intra prediction applies to the block, based on mode information of the block, the mode information of the block comprising at least one of the following: a reference line index of the block, an intra prediction mode of the block, and a size of the block; and in response to determining that the offset refinement for intra prediction applies to the block, performing, by the device, the offset refinement to generate an intra predictor for intra prediction of the block.
H04N 19/593 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
H04N 19/11 - Selection of coding mode or of prediction mode among a plurality of spatial predictive coding modes
H04N 19/13 - Adaptive entropy coding, e.g. adaptive variable length coding [AVLC] or context adaptive binary arithmetic coding [CABAC]
H04N 19/136 - Incoming video signal characteristics or properties
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
63.
SYSTEMS AND METHODS FOR CHROMA MODE CODING IN RECURSIVE INTRA/INTER REGION
An example method of video coding includes receiving a video bitstream including a plurality of blocks; identifying, based on a first indicator in the video bitstream, a coding region that comprises two or more blocks of the plurality of blocks, where each block in the coding region is encoded in an intra prediction mode. The method further includes identifying, for a chroma block in the coding region, an angular intra prediction mode of a collocated luma block; and reconstructing the two or more blocks of the coding region using the angular intra prediction mode of the collocated luma block as a reference for the chroma block.
H04N 19/169 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
H04N 19/119 - Adaptive subdivision aspects e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
H04N 19/12 - Selection from among a plurality of transforms or standards, e.g. selection between discrete cosine transform [DCT] and sub-band transform or selection between H.263 and H.264
H04N 19/44 - Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
64.
SYSTEMS AND METHODS FOR ENTROPY CODING FOR RECURSIVE INTRA REGION PARTITIONING
An example method of video coding includes receiving a video bitstream including a plurality of coding blocks; identifying a coding region that includes two or more coding blocks of the plurality of coding blocks that are encoded in a first prediction mode. The method also includes entropy decoding a signaled flag indicating a prediction mode for the coding region, the entropy decoding using coded information including one or more of: previous instances of the signaled flag, a block size of a current coding block, a block size group of the current coding block, respective block sizes of a set of neighboring coding blocks, respective block size groups of the set of neighboring coding blocks, and signaled flags for the set of neighboring coding blocks; and reconstructing the two or more coding blocks according to a value of the signaled flag for the coding region.
H04N 19/169 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
H04N 19/119 - Adaptive subdivision aspects e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
H04N 19/12 - Selection from among a plurality of transforms or standards, e.g. selection between discrete cosine transform [DCT] and sub-band transform or selection between H.263 and H.264
H04N 19/44 - Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
65.
SYSTEMS AND METHODS FOR TRANSFORM PARTITIONING IN A RECURSIVE INTRA REGION
An example method of video coding includes receiving a video bitstream including a plurality of frames; identifying a coding region in a frame of the plurality of frames, the coding region including two or more blocks of the plurality of blocks, where each block in the coding region is encoded in an intra prediction mode. The method includes obtaining a set of luma transform blocks according to a first transform partitioning for luma blocks in the coding region and obtaining a set of chroma transform blocks according to a second transform partitioning for chroma blocks in the coding region, where the second transform partitioning is different than the first transform partitioning. The method further includes reconstructing the two or more blocks of the coding region using respective prediction blocks, the set of luma transform blocks, and the set of chroma transform blocks.
H04N 19/169 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
H04N 19/119 - Adaptive subdivision aspects e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
H04N 19/12 - Selection from among a plurality of transforms or standards, e.g. selection between discrete cosine transform [DCT] and sub-band transform or selection between H.263 and H.264
H04N 19/44 - Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
Processing circuitry performs a dequantization on a first quantized block that is decoded from the coded information of the current block to generate a dequantized block. The processing circuitry performs an inverse transform on the dequantized block to obtain an inverse transform block, combines the first predictor of the current block with the inverse transform block of the dequantized block to generate an intermediate reconstruction of the current block, generates a second predictor of the current block based on one or more neighboring blocks of the current block, performs a transform on a difference of the intermediate reconstruction of the current block and the second predictor of the current block to generate a transform block, performs a quantization on the transform block to generate a second quantized block, and determines at least a quantization shifting offset based on the dequantized block and the second quantized block.
H04N 19/126 - Details of normalisation or weighting functions, e.g. normalisation matrices or variable uniform quantisers
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/593 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
67.
SYSTEMS AND METHODS FOR END OF BLOCK CODING FOR 2D COEFFICIENTS BLOCK WITH 1D TRANSFORMS
The various implementations described herein include methods and systems for coding video. In one aspect, a method includes receiving a video bitstream comprising a plurality of blocks, including a current block. The method includes parsing, from the video bitstream, a plurality of one-dimensional (ID) transform coefficients corresponding to the current block. The method includes parsing, from the video bitstream, a maximum end-of- block (BOB) value for the plurality of ID transform coefficients. The method includes deriving a two-dimensional (2D) transform block using the maximum BOB value and the plurality of ID transform coefficients. The method also includes reconstructing the current block using the derived 2D transform block.
H04N 19/12 - Selection from among a plurality of transforms or standards, e.g. selection between discrete cosine transform [DCT] and sub-band transform or selection between H.263 and H.264
H04N 19/136 - Incoming video signal characteristics or properties
H04N 19/18 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a set of transform coefficients
H04N 19/60 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
H04N 19/40 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video transcoding, i.e. partial or full decoding of a coded input stream followed by re-encoding of the decoded output stream
H04N 19/61 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
68.
SYSTEMS AND METHODS FOR LOOP FILTERING FOR MULTI-VIEW CODING
The various implementations described herein include methods and systems for coding video. In one aspect, a method includes receiving a multi-view video bitstream comprising a plurality of pictures. The plurality of pictures includes a first picture corresponding to a first view and a second picture corresponding to a second view. The method includes determining, based on a first indicator in the multi-view video bitstream, whether loop filtering parameters for the first picture corresponding to the first view and the second picture corresponding to the second view are signaled jointly or separately. The method includes, in accordance with the first indicator indicating that the loop filtering parameters are signaled jointly, performing a first loop filtering process on the first picture and a second loop filtering process on the second picture using a shared set of loop filtering parameters.
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
H04N 19/00 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
H04N 19/117 - Filters, e.g. for pre-processing or post-processing
69.
DECODER SIDE QUANTIZATION SHIFTING OFFSET PREDICTION
An apparatus includes processing circuitry configured to receive a bitstream that comprises coded information of a current block, and determine a plurality of hypotheses for a decoder side quantization shifting offset prediction, a hypothesis in the plurality of hypotheses corresponds to a potential quantization shifting offset setting in a transform domain of the current block. The processing circuitry calculates cost values respectively associated with the plurality of hypotheses, select a specific hypothesis from the plurality of hypotheses according to the cost values, determine one or more quantization shifting offsets for transform coefficients in the transform domain based on the specific hypothesis, reconstruct the transform coefficients based on the one or more quantization shifting offsets, calculate residuals in a spatial domain of the current block based on the transform coefficients in the transform domain, and reconstruct the current block according to the residuals in the spatial domain.
H04N 19/147 - Data rate or code amount at the encoder output according to rate distortion criteria
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/40 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video transcoding, i.e. partial or full decoding of a coded input stream followed by re-encoding of the decoded output stream
70.
ADAPTIVE CLIPPING IN MODELS PARAMETERS DERIVATIONS METHODS FOR VIDEO COMPRESSION
A method includes receiving a bitstream of coded information for pictures. The coded information indicates a prediction of a current block using a model based prediction technique, the model based prediction technique generates a prediction sample of the current block based on a model with one or more reconstructed samples of a reference block being input to the model, the model includes one or more parameters derived based on a current template of the current block and a reference template of the reference block. The method also includes performing at least a clipping operation on at least one of the current template and the reference template to obtain clipped template samples; deriving one or more parameter values of the one or more parameters of the model according to the clipped template samples; and generating at least a prediction sample of the current block by the model.
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/117 - Filters, e.g. for pre-processing or post-processing
H04N 19/82 - Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation involving filtering within a prediction loop
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
71.
SYSTEMS AND METHODS FOR DECODER-SIDE MOTION VECTOR REFINEMENT
An example method of video coding includes receiving a video bitstream comprising a plurality of blocks. The method also includes deriving a set of subblock motion vectors for a current subblock of a current block of the plurality of blocks and deriving a set of refined subblock motion vectors for the current subblock using bilateral matching that is based on one or more samples outside of the current subblock. The method further includes reconstructing the current subblock using the derived set of refined subblock motion vectors.
H04N 19/139 - Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
A method includes receiving, by a user interface layer in a client terminal, application code input by a user, processing the input application code by a client-as-a-service layer in the client terminal, and forwarding the input application code to be stored in a cloud-based repository. The method further includes synchronizing application code stored in a cache with application code stored in the cloud-based repository, and performing a runtime remodel on a runtime system based on the input application code.
H04L 67/60 - Scheduling or organising the servicing of application requests, e.g. requests for application data transmissions using the analysis and optimisation of the required network resources
H04L 67/10 - Protocols in which an application is distributed across nodes in the network
G06F 16/178 - Techniques for file synchronisation in file systems
H04L 67/00 - Network arrangements or protocols for supporting network services or applications
Aspects of the disclosure includes methods and apparatuses for video decoding and encoding and a method of processing visual media data. The method for video decoding includes receiving coded information in a bitstream indicating that a current block is predicted based on a combination of a plurality of intra prediction modes. The method includes determining a plurality of intra predictions of the current block based on the respective intra prediction modes, determining a fused prediction of the current block based on a weighted summation of the plurality of intra predictions where the weighted summation is according to respective weights associated with the plurality of intra predictions, and reconstructing the current block based on the fused prediction. Each weight is based on one of a plurality of weighting functions that depends on a sample location (x, y) and the intra prediction mode of the intra prediction associated with the respective weight.
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/11 - Selection of coding mode or of prediction mode among a plurality of spatial predictive coding modes
H04N 19/50 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
74.
SYSTEMS AND METHODS FOR TRANSFORM PARTITIONING IN A RECURSIVE INTRA REGION
An example method of video coding includes receiving a video bitstream including a plurality of frames; identifying a coding region in a frame of the plurality of frames, the coding region including two or more blocks of the plurality of blocks, where each block in the coding region is encoded in an intra prediction mode. The method includes obtaining a set of luma transform blocks according to a first transform partitioning for luma blocks in the coding region and obtaining a set of chroma transform blocks according to a second transform partitioning for chroma blocks in the coding region, where the second transform partitioning is different than the first transform partitioning. The method further includes reconstructing the two or more blocks of the coding region using respective prediction blocks, the set of luma transform blocks, and the set of chroma transform blocks.
H04N 19/44 - Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
H04N 19/119 - Adaptive subdivision aspects e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
75.
SYSTEMS AND METHODS FOR IMPLICIT DERIVATION IN A RECURSIVE INTRA REGION
An example method of video coding includes receiving a video bitstream including a plurality of coding blocks; identifying, based on a first indicator in the video bitstream, a coding region that comprises two or more blocks of the plurality of coding blocks, where at least one block in the coding region is encoded in an intra prediction mode and at least one block in the coding region is encoded in a non-intra prediction mode. For a last block in the coding region, when all preceding blocks in the coding region are encoded in the intra prediction mode, the method includes inferring a prediction mode of the last block as being an inter prediction mode without receiving an inter prediction mode indicator for the last block; and reconstructing the last block using the inferred prediction mode.
H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
76.
SYSTEMS AND METHODS FOR END OF BLOCK CODING FOR 2D COEFFICIENTS BLOCK WITH 1D TRANSFORMS
The various implementations described herein include methods and systems for coding video. In one aspect, a method includes receiving a video bitstream comprising a plurality of blocks, including a current block. The method includes parsing, from the video bitstream, a plurality of one-dimensional (1D) transform coefficients corresponding to the current block. The method includes parsing, from the video bitstream, a maximum end-of-block (EOB) value for the plurality of 1D transform coefficients. The method includes deriving a two-dimensional (2D) transform block using the maximum EOB value and the plurality of 1D transform coefficients. The method also includes reconstructing the current block using the derived 2D transform block.
H04N 19/44 - Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/18 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a set of transform coefficients
77.
PREDICTION MODE SELECTION IN POLYGON MESH COMPRESSION
A bitstream that includes attribute information of a plurality of attributes in a mesh is received. One or more prediction modes are determined from a plurality of candidate prediction modes for a current attribute of the plurality of attributes in the mesh based on one of prediction mode priorities and prediction mode accuracies of the one or more prediction modes in the plurality of candidate prediction modes. A prediction value of the current attribute in the mesh is determined based on the attribute information of the current attribute and the one or more prediction modes. The current attribute is reconstructed based on the prediction value of the current attribute.
A method and apparatus comprising computer code configured to cause a processor or processors to obtain, from a bitstream, a mesh representing an encoded volumetric data of at least one three-dimensional (3D) visual content, partition a plurality of vertices of the mesh into a plurality of groups by determining a face normal for each face in the mesh, classifying the groups based on their orientation to the face normal, and merging adjacent faces that share an edge between the plurality of groups, and decode the encoded volumetric data based on the plurality of groups.
H04N 19/14 - Coding unit complexity, e.g. amount of activity or edge presence estimation
H04N 19/119 - Adaptive subdivision aspects e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
H04N 19/169 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
In a method, a bitstream of a mesh that includes a plurality of vertices is received. When the mesh is symmetric with respect to a symmetry plane, a first group of vertices of the mesh that is positioned at a first side of the symmetry plane is reconstructed. A second group of vertices of the mesh that is positioned at a second side of the symmetry plane is reconstructed by reflecting the reconstructed first group of vertices with respect to the symmetry plane.
An example method of video coding includes receiving a video bitstream including a plurality of coding blocks; identifying a coding region that includes two or more coding blocks of the plurality of coding blocks that are encoded in a first prediction mode. The method also includes entropy decoding a signaled flag indicating a prediction mode for the coding region, the entropy decoding using coded information including one or more of: previous instances of the signaled flag, a block size of a current coding block, a block size group of the current coding block, respective block sizes of a set of neighboring coding blocks, respective block size groups of the set of neighboring coding blocks, and signaled flags for the set of neighboring coding blocks; and reconstructing the two or more coding blocks according to a value of the signaled flag for the coding region.
H04N 19/44 - Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
H04N 19/13 - Adaptive entropy coding, e.g. adaptive variable length coding [AVLC] or context adaptive binary arithmetic coding [CABAC]
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/91 - Entropy coding, e.g. variable length coding [VLC] or arithmetic coding
81.
SYSTEMS AND METHODS FOR CHROMA MODE CODING IN RECURSIVE INTRA/INTER REGION
An example method of video coding includes receiving a video bitstream including a plurality of blocks; identifying, based on a first indicator in the video bitstream, a coding region that comprises two or more blocks of the plurality of blocks, where each block in the coding region is encoded in an intra prediction mode. The method further includes identifying, for a chroma block in the coding region, an angular intra prediction mode of a collocated luma block; and reconstructing the two or more blocks of the coding region using the angular intra prediction mode of the collocated luma block as a reference for the chroma block.
H04N 19/11 - Selection of coding mode or of prediction mode among a plurality of spatial predictive coding modes
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/107 - Selection of coding mode or of prediction mode between spatial and temporal predictive coding, e.g. picture refresh
H04N 19/139 - Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
82.
SYSTEMS AND METHODS FOR APPLYING NON-SEPARABLE TRANSFORMS ON INTER PREDICTION RESIDUALS
The various implementations described herein include methods and systems for coding video. In one aspect, a method includes receiving a video bitstream that includes a set of inter mode encoded blocks and a corresponding set of transform coefficients. The method includes deriving a set of inter mode residual blocks from the set of transform coefficients. The method includes determining, according to a value of a first indicator in the video bitstream, whether one or more non-separable transform kernels are to be applied to the set of inter mode residual blocks. The method includes applying a first non-separable transform kernel when the indicator has a first value, and forgoing applying the first non-separable transform kernels when the indicator has a second value. The method also includes reconstructing a set of video blocks using the set of inter mode residual blocks and a corresponding set of prediction blocks.
H04N 19/60 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/18 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a set of transform coefficients
A method includes receiving, by a user interface layer in a client terminal, application code input by a user, processing the input application code by a client-as-a-service layer in the client terminal, and forwarding the input application code to be stored in a cloud-based repository. The method further includes synchronizing application code stored in a cache with application code stored in the cloud-based repository, and performing a runtime remodel on a runtime system based on the input application code.
An example method of video coding includes receiving a video bitstream comprising a plurality of blocks. The method also includes deriving a set of subblock motion vectors for a current subblock of a current block of the plurality of blocks and deriving a set of refined subblock motion vectors for the current subblock using bilateral matching that is based on one or more samples outside of the current subblock. The method further includes reconstructing the current subblock using the derived set of refined subblock motion vectors.
H04N 19/52 - Processing of motion vectors by encoding by predictive encoding
H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
H04N 19/139 - Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
H04N 19/157 - Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
H04N 19/167 - Position within a video image, e.g. region of interest [ROI]
H04N 19/172 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/80 - Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation
85.
SYSTEMS AND METHODS FOR LOOP FILTERING FOR MULTI-VIEW CODING
The various implementations described herein include methods and systems for coding video. In one aspect, a method includes receiving a multi-view video bitstream comprising a plurality of pictures. The plurality of pictures includes a first picture corresponding to a first view and a second picture corresponding to a second view. The method includes determining, based on a first indicator in the multi-view video bitstream, whether loop filtering parameters for the first picture corresponding to the first view and the second picture corresponding to the second view are signaled jointly or separately. The method includes, in accordance with the first indicator indicating that the loop filtering parameters are signaled jointly, performing a first loop filtering process on the first picture and a second loop filtering process on the second picture using a shared set of loop filtering parameters.
H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
H04N 19/117 - Filters, e.g. for pre-processing or post-processing
H04N 19/172 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
H04N 19/82 - Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation involving filtering within a prediction loop
86.
ADAPTIVE CLIPPING IN MODELS PARAMETERS DERIVATIONS METHODS FOR VIDEO COMPRESSION
A method includes receiving a bitstream of coded information for pictures. The coded information indicates a prediction of a current block using a model based prediction technique, the model based prediction technique generates a prediction sample of the current block based on a model with one or more reconstructed samples of a reference block being input to the model, the model includes one or more parameters derived based on a current template of the current block and a reference template of the reference block. The method also includes performing at least a clipping operation on at least one of the current template and the reference template to obtain clipped template samples; deriving one or more parameter values of the one or more parameters of the model according to the clipped template samples; and generating at least a prediction sample of the current block by the model.
H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
87.
SYSTEMS AND METHODS FOR CROSS-VIEW MOTION VECTOR PREDICTION
The various implementations described herein include methods and systems for coding video. In one aspect, a method includes receiving a multi-view video bitstream that includes a first block in a first frame corresponding to a first view and a second block in a second frame corresponding to a second view. The method identifies a first set of reference frames in the first view for the first block. The method obtains motion vectors corresponding to a second set of reference frames in the second view for the second block. In accordance with a determination that the first set of reference frames share a display time with the second set of reference frames, the method derives a motion vector predictor (MVP) for the first block corresponding to the first view using the set of motion vectors corresponding to the second view, and decodes the first block using the derived MVP.
H04N 13/00 - Stereoscopic video systemsMulti-view video systemsDetails thereof
H04N 19/52 - Processing of motion vectors by encoding by predictive encoding
H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/51 - Motion estimation or motion compensation
Embodiments of the present disclosure are directed to a method to process immersive media, including receiving scene graph information of a first scene in a first scene graph format to be converted into a second scene graph format. The method may also include obtaining a metadata framework to preserve scene information stored in a scene graph during a process of scene graph translation, the metadata framework comprising a plurality of subsystems; converting the first scene into the second scene graph format using the metadata framework; and rendering the first scene in the second scene graph format based on the converting. The plurality of subsystems may include a subsystem comprising information associated with geometric assets of the first scene; a subsystem comprising information associated with animation of one or more assets in the first scene; and a subsystem comprising information associated with a logical sequence of data in the first scene.
H04N 21/234 - Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
H04N 21/44 - Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
H04N 21/84 - Generation or processing of descriptive data, e.g. content descriptors
H04N 21/8543 - Content authoring using a description language, e.g. MHEG [Multimedia and Hypermedia information coding Expert Group] or XML [eXtensible Markup Language]
89.
Adaptive Codebook for Neural Network-Based Audio Codec
This disclosure relates generally to audio coding and particularly to methods and systems for audio coding based on neural networks. In particular, feature vectors generated by a neural network audio encoder may be quantized using adaptive codebooks and/or grouped codebooks. Correspondingly, the encoded bitstream may be processed via a dequantization process using the adaptive codebooks and/or grouped codebooks. The adaptive codebooks or grouped code books may be selected to preserve a maximum bitrate and potentially increase coding efficiency
G10L 19/032 - Quantisation or dequantisation of spectral components
G10L 19/00 - Speech or audio signal analysis-synthesis techniques for redundancy reduction, e.g. in vocodersCoding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
Aspects of the disclosure includes methods and apparatuses for video decoding and encoding and a method of processing visual media data. The method for video decoding includes receiving coded information in a bitstream indicating that a current block is predicted based on a combination of a plurality of intra prediction modes. The method includes determining a plurality of intra predictions of the current block based on the respective intra prediction modes, determining a fused prediction of the current block based on a weighted summation of the plurality of intra predictions where the weighted summation is according to respective weights associated with the plurality of intra predictions, and reconstructing the current block based on the fused prediction. Each weight is based on one of a plurality of weighting functions that depends on a sample location (x, y) and the intra prediction mode of the intra prediction associated with the respective weight.
H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
H04N 19/119 - Adaptive subdivision aspects e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
91.
METHOD AND APPARATUS FOR NEURAL SPATIAL SPEECH CODING FOR MULTI-CHANNEL AUDIO
A method performed for performing neural spatial audio coding, comprises: receiving an audio signal comprising a plurality of channels; selecting a channel from the plurality of channels as a reference channel; performing a STFT on the reference channel to generate a frequency domain reference channel; inputting the frequency domain reference channel into a first codec; performing the STFT on the plurality of channels minus the channel selected as the reference channel to generate a spatial covariance matrix; inputting the spatial covariance matrix and the frequency domain reference channel into a second codec; reconstructing the audio signal based on an output of the first codec and an output of the second codec to generate a reconstructed audio signal; and training the first and second codecs.
G10L 19/008 - Multichannel audio signal coding or decoding using interchannel correlation to reduce redundancy, e.g. joint-stereo, intensity-coding or matrixing
G10L 19/02 - Speech or audio signal analysis-synthesis techniques for redundancy reduction, e.g. in vocodersCoding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using spectral analysis, e.g. transform vocoders or subband vocoders
G10L 19/032 - Quantisation or dequantisation of spectral components
G10L 25/18 - Speech or voice analysis techniques not restricted to a single one of groups characterised by the type of extracted parameters the extracted parameters being spectral information of each sub-band
G10L 25/30 - Speech or voice analysis techniques not restricted to a single one of groups characterised by the analysis technique using neural networks
92.
METHOD AND APPARATUS FOR NEURAL SPATIAL SPEECH CODING FOR MULTI-CHANNEL AUDIO
A method performed for performing neural spatial audio coding, comprises: receiving an audio signal comprising a plurality of channels; selecting a channel from the plurality of channels as a reference channel; performing a STFT on the reference channel to generate a frequency domain reference channel; inputting the frequency domain reference channel into a first codec; performing the STFT on the plurality of channels minus the channel selected as the reference channel to generate a spatial covariance matrix; inputting the spatial covariance matrix and the frequency domain reference channel into a second codec; reconstructing the audio signal based on an output of the first codec and an output of the second codec to generate a reconstructed audio signal; and training the first and second codecs.
G10L 19/008 - Multichannel audio signal coding or decoding using interchannel correlation to reduce redundancy, e.g. joint-stereo, intensity-coding or matrixing
G10L 19/02 - Speech or audio signal analysis-synthesis techniques for redundancy reduction, e.g. in vocodersCoding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using spectral analysis, e.g. transform vocoders or subband vocoders
G10L 21/00 - Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
An example method of video coding includes receiving a video bitstream for a plurality of encoded pictures. The method also includes determining, based on a first indicator, whether a frame-level frame interpolation mode is enabled for an encoded picture and determining, based on a second indicator, whether a prediction filtering mode is enabled for the encoded picture. The method further includes determining, based on the first indicator indicating that the frame-level frame interpolation mode is enabled and the second indicator indicating that the prediction filtering mode is enabled, that the video bitstream includes a third indicator that indicates whether one or more quantization parameters for the encoded picture are signaled in the video bitstream.
An example method includes receiving a multi-view video bitstream comprising a plurality of blocks. The plurality of blocks includes a first block in a first frame corresponding to a first view and a second block in a second frame corresponding to a second view. The method includes determining, based on a first indicator in the multi-view video bitstream, whether block partitioning modes for the first block corresponding to the first view and the second block corresponding to the second view are signaled jointly or separately. The method includes, in accordance with the first indicator indicating that the block partitioning modes are signaled jointly, applying a same block partitioning mode to the first block and the second block in accordance with a second indicator.
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
H04N 19/00 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
H04N 19/172 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
H04N 19/46 - Embedding additional information in the video signal during the compression process
95.
ZERO BYTE CODING OF ATTRIBUTE CONNECTIVITY IN POLYGON MESHES
Some aspects of the disclosure provide a method of mesh processing. The method includes receiving a bitstream of coded information of a mesh, the mesh includes a plurality of three-dimensional (3D) vertices in a 3D space, and at least a non-position attribute. The coded information includes a position connectivity of the plurality of 3D vertices in the 3D space. The method also includes determining whether the coded information of the mesh indicates that a non-position attribute connectivity of the non-position attribute corresponds to at least an extreme case, and determining the non-position attribute connectivity according to the position connectivity of the plurality of 3D vertices when the non-position attribute connectivity corresponds to the extreme case.
H04N 19/20 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video object coding
H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
96.
EFFICIENT CODING OF ATTRIBUTES IN POLYGON MESH COMPRESSION
A bitstream that includes attribute information of a plurality of UV vertices corresponding to a mesh is received. A seam vertex queue that includes one or more UV vertices of the plurality of UV vertices is generated. Each of the one or more UV vertices in the seam vertex queue belongs to a respective UV vertex group of a plurality of seam UV vertex groups. Each of the plurality of seam UV vertex groups corresponds to a same three-dimensional (3D) vertex of the mesh in a 3D space. An initial UV vertex of the mesh is determined based on whether one of the one or more UV vertices in the seam vertex queue includes an unvisited UV vertex. UV coordinates of the mesh are reconstructed based on the initial UV vertex.
Aspects of the disclosure includes methods and apparatuses for coding a mesh. A method for decoding a mesh includes: receiving coded information including a syntax element indicating one of an across-parallelogram prediction and a reflection prediction to predict a position of one of a current vertex of the mesh and a current point of a two-dimensional (2D) map that is associated with the mesh, determining the one of the across-parallelogram prediction and the reflection prediction to predict the position of the one of the current vertex of the mesh and the current point of the 2D map based on the syntax element, and predicting the position of the one of the current vertex of the mesh and the current point of the 2D map using the determined one of the across-parallelogram prediction and the reflection prediction.
Embodiments of the present disclosure are directed to a method to process immersive media, including receiving scene graph information of a first scene in a first scene graph format to be converted into a second scene graph format. The method may also include obtaining a metadata framework to preserve scene information stored in a scene graph during a process of scene graph translation, the metadata framework comprising a plurality of subsystems; converting the first scene into the second scene graph format using the metadata framework; and rendering the first scene in the second scene graph format based on the converting. The plurality of subsystems may include a subsystem comprising information associated with geometric assets of the first scene; a subsystem comprising information associated with animation of one or more assets in the first scene; and a subsystem comprising information associated with a logical sequence of data in the first scene.
Aspects of the disclosure provide a method and an apparatus including processing circuitry that determines, based on a syntax element in a coded video bitstream, that a current block including a plurality of subblocks is coded in a subblock-based temporal motion vector prediction (SbTMVP) mode. Motion vector offset (MVO) information indicating an MVO is received. The MVO indicates a motion offset of a displacement vector (DV) used to adjust a location of a collocated block in a collocated reference picture. An updated DV of the current block is determined based on the DV and the MVO. SbTMVP information of a respective subblock in the plurality of subblocks is derived based on motion information of a corresponding subblock in the collocated block indicated by the updated DV. The plurality of subblocks in the SbTMVP mode is reconstructed based on the SbTMVP information of the subblock in the plurality of subblocks.
H04N 19/137 - Motion inside a coding unit, e.g. average field, frame or block difference
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
100.
SYSTEMS AND METHODS FOR IMPROVED RECURSIVE INTRA REGION PARTITIONING
An example method of video coding includes receiving a video bitstream for a plurality of coding blocks. The method also includes identifying, based on a first indicator in the video bitstream, a coding region that includes two or more coding blocks of the plurality of coding blocks, where each coding block in the coding region is encoded in an intra prediction mode. The method further includes applying a first partitioning for luma blocks in the coding region, applying a second partitioning for chroma blocks in the coding region, where the second partitioning is different from the first partitioning, and reconstructing the two or more coding blocks of the coding region using the first partitioning and the second partitioning.
H04N 19/169 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
H04N 19/119 - Adaptive subdivision aspects e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
H04N 19/12 - Selection from among a plurality of transforms or standards, e.g. selection between discrete cosine transform [DCT] and sub-band transform or selection between H.263 and H.264
H04N 19/44 - Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder