Настройки

Укажите год
-

Небесная энциклопедия

Космические корабли и станции, автоматические КА и методы их проектирования, бортовые комплексы управления, системы и средства жизнеобеспечения, особенности технологии производства ракетно-космических систем

Подробнее
-

Мониторинг СМИ

Мониторинг СМИ и социальных сетей. Сканирование интернета, новостных сайтов, специализированных контентных площадок на базе мессенджеров. Гибкие настройки фильтров и первоначальных источников.

Подробнее

Форма поиска

Поддерживает ввод нескольких поисковых фраз (по одной на строку). При поиске обеспечивает поддержку морфологии русского и английского языка
Ведите корректный номера.
Ведите корректный номера.
Ведите корректный номера.
Ведите корректный номера.
Укажите год
Укажите год

Применить Всего найдено 515. Отображено 160.
23-01-2018 дата публикации

Encoding and decoding architecture of checkerboard multiplexed image data

Номер: US0009877045B2

A device includes a coder or a codec configured for interleaved image data utilizing diamond shaped blocks for motion estimation and/or motion compensation and utilizing square or orthogonal transforms of residual data. In various embodiments, the decoder may be configured, among others, to perform de-blocking on edges of the diamond shaped blocks and/or data padding at boundaries of the image data.

Подробнее
30-05-2017 дата публикации

Reduced complexity motion compensated temporal processing

Номер: US0009667964B2

A method and system for reduced complexity motion compensated temporal processing for pre-analysis purposes. The processing complexity is reduced by reducing the number of pictures processed for analysis, reducing the number of references used for analysis, and/or spatially subsampling input pictures.

Подробнее
17-10-2012 дата публикации

Filtering for image and video enhancement using asymmetric samples

Номер: CN102742269A
Принадлежит:

Methods and devices for filtering lower quality images or sequences of images with higher quality images or sequences of images are described.

Подробнее
27-04-2017 дата публикации

SPATIOTEMPORAL PREDICTION FOR BIDIRECTIONALLY PREDICTIVE (B) PICTURES AND MOTION VECTOR PREDICTION FOR MULTI-PICTURE REFERENCE MOTION COMPENSATION

Номер: US20170118488A1
Принадлежит: Microsoft Technology Licensing, LLC

Several improvements for use with Bidirectionally Predictive (B) pictures within a video sequence are provided. In certain improvements Direct Mode encoding and/or Motion Vector Prediction are enhanced using spatial prediction techniques. In other improvements Motion Vector prediction includes temporal distance and subblock information, for example, for more accurate prediction. Such improvements and other presented herein significantly improve the performance of any applicable video coding system/logic.

Подробнее
08-08-2017 дата публикации

Directed interpolation and data post-processing

Номер: US0009729899B2

An encoding device evaluates a plurality of processing and/or post-processing algorithms and/or methods to be applied to a video stream, and signals a selected method, algorithm, class or category of methods/algorithms either in an encoded bitstream or as side information related to the encoded bitstream. A decoding device or post-processor utilizes the signaled algorithm or selects an algorithm/method based on the signaled method or algorithm. The selection is based, for example, on availability of the algorithm/method at the decoder/post-processor and/or cost of implementation. The video stream may comprise, for example, downsampled multiplexed stereoscopic images and the selected algorithm may include any of upconversion and/or error correction techniques that contribute to a restoration of the downsampled images.

Подробнее
12-06-2013 дата публикации

Systems and methods for multi-layered image and video delivery using reference processing signals

Номер: CN103155568A
Принадлежит:

The present disclosure describes a system and method for multi-layered image and video delivery using reference processing signal. A multi-layered encoder and a multi-layered decoder, both comprising at least one enhancement layer processor that uses reference processing and a base layer processor are provided. A multi-layered encoder and a multi-layered decoder that use reference processing and are capable of frame-compatible 3D video delivery are also described.

Подробнее
22-11-2016 дата публикации

Filtering for image and video enhancement using asymmetric samples

Номер: US0009503757B2

Filtering lower quality images or sequences of images with higher quality images or sequences of images. The filtering is utilized in a post-process, decoded, or encoded images including multiple sets of images that are filtered and/or combined. Different image features including, for example, quality, frequency characteristics, temporal resolution, spatial resolution, number of views, or bit-depth are present in the images. In one embodiment, the filtering comprises a multi-hypothesis filtering and the confidence value comprises a difference between the filtered sample and samples lying within a filter support. The post processes images are then stored or distributed.

Подробнее
16-08-2016 дата публикации

Coding and decoding of interleaved image data

Номер: US0009420311B2

Sampled data is packaged in checkerboard format for encoding and decoding. The sampled data may be quincunx sampled multi-image video data (e.g., 3D video or a multi-program stream), and the data may also be divided into sub-images of each image which are then multiplexed, or interleaved, in frames of a video stream to be encoded and then decoded using a standardized video encoder. A system for viewing may utilize a standard video decoder and a formatting device that de-interleaves the decoded sub-images of each frame reformats the images for a display device. A 3D video may be encoded using a most advantageous interleaving format such that a preferred quality and compression ratio is reached. In one embodiment, the invention includes a display device that accepts data in multiple formats.

Подробнее
03-10-2012 дата публикации

Buffered adaptive filters

Номер: CN102714727A
Принадлежит:

A buffer is established in a video encoder and/or a video decoder. An adaptive filter is buffered in the established buffer. An input video signal is coded using the buffered filter. The filter buffer may be managed, e.g., as to buffer size or capacity, filter placement within the buffer, and buffered default filters. Adaptive filters may be signaled and buffered for multiple reference pictures.

Подробнее
25-10-2016 дата публикации

Complexity allocation for video and image coding applications

Номер: US0009479786B2

Coding method complexity types are assigned to video for combinations of coding decision levels associated with a sequence of pictures, a group of pictures, a scene, a region of the scene or picture, a block or macroblock. Multiple coding units are encoded as a function of complexity and determined from the coding decision levels. A first coding decision level is associated with determining a first complexity and a first coding method type for the sequence of pictures. A second coding decision level is associated with determining a second complexity and a second coding method type for the pictures or scene. The picture or sequence includes the coding units. A third coding decision level is associated with determining a third complexity and a third coding method type for the coding units.

Подробнее
22-06-2017 дата публикации

Rate Control for Video Splicing Applications

Номер: US20170180735A1

Rate control techniques are provided for encoding an input video sequence into a compressed coded bitstream with multiple coding passes. The final coding pass may comprise final splices with non-overlapping frames that do not extend into neighboring final splices. A final splice in the final coding pass may correspond to at least one non-final splice in a non-final coding pass. A non-final splice may have overlapping frames that extend into neighboring final splices in the final coding pass. The overlapping frames in the non-final splice may be used to derive complexity information about the neighboring final splices. The complexity information about the neighboring final splices, as derived from the overlapping frames, may be used to allocate or improve rate control related budgets in encoding the final splice into the compressed coded bitstream in the final coding pass.

Подробнее
16-03-2017 дата публикации

CHROMA QUANTIZATION IN VIDEO CODING

Номер: US20170078667A1
Принадлежит:

A method of signaling additional chroma QP offset values that are specific to quantization groups is provided, in which each quantization group explicitly specifies its own set of chroma QP offset values. Alternatively, a table of possible sets of chroma QP offset values is specified in the header area of the picture, and each quantization group uses an index to select an entry from the table for determining its own set of chroma QP offset values. The quantization group specific chroma QP offset values are then used to determine the chroma QP values for blocks within the quantization group in addition to chroma QP offset values already specified for higher levels of the video coding hierarchy. 127-. (canceled)28. A non-transitory computer readable medium storing a program that is executable by a processing unit , the program comprising sets of instructions for:identifying two or more initial sets of chroma quantization parameter (QP) offset values at two or more levels of a video coding hierarchy, each initial set of chroma QP offset values for specifying chroma QPs of video units encompassed by one level of the video coding hierarchy;identifying an additional set of chroma QP offset values for a quantization group comprising a plurality of video units; andcomputing, for the plurality of video units, a set of chroma QP values by adding (i) the initial sets of chroma QP offset values that were identified for the plurality of video units and (ii) the additional set of chroma QP offset values that was identified for the quantization group.29. The non-transitory computer readable medium of claim 28 , wherein an identified initial set of chroma QP offset values specified at a particular level of the video coding hierarchy are for video units that are encompassed by the particular level of the video coding hierarchy.30. The method of further comprising identifying a luma QP value for the plurality of video units.31. The method of claim 30 , wherein computing the set of ...

Подробнее
05-06-2013 дата публикации

Optimized filter selection for reference picture processing

Номер: CN103141099A
Принадлежит:

Reference processing may be used in a video encoder or decoder to derive reference pictures that are better correlated with a source image to be encoded or decoded, which generally yields better coding efficiency. Methods for filter selection for a reference processing unit adapted for use in a video codec system are discussed. Specifically, methods for filter selection based on performing motion estimation and obtaining distortion/cost information by comparing reference pictures, either processed or non-processed, with the source image to be encoded are discussed.

Подробнее
11-10-2016 дата публикации

Systems and methods for multi-layered image and video delivery using reference processing signals

Номер: US0009467689B2

The present disclosure describes a system and method for multi-layered image and video delivery using reference processing signal. A multi-layered encoder and a multi-layered decoder, both comprising at least one enhancement layer processor that uses reference processing and a base layer processor are provided. A multi-layered encoder and a multi-layered decoder that use reference processing and are capable of frame-compatible 3D video delivery are also described.

Подробнее
13-09-2016 дата публикации

Overlapped block disparity estimation and compensation architecture

Номер: US0009445121B2

Overlapped block disparity estimation and compensation is described. Compensating for images with overlapped block disparity compensation (OBDC) involves determining if OBDC is enabled in a video bit stream, and determining if OBDC is enabled for one or more macroblocks that neighbor a first macroblock within the video bit stream. The neighboring macroblocks may be transform coded. If OBDC is enabled in the video bit stream and for the one or more neighboring macroblocks, predictions may be made for a region of the first macroblock that has an edge adjacent with the neighboring macroblocks. OBDC can be causally applied. Disparity compensation parameters or modes may be shared amongst views or layers. A variety of predictions may be used with causally-applied OBDC.

Подробнее
23-01-2018 дата публикации

Coding and decoding of interleaved image data

Номер: US0009877047B2

Sampled data is packaged in checkerboard format for encoding and decoding. The sampled data may be quincunx sampled multi-image video data (e.g., 3D video or a multi-program stream), and the data may also be divided into sub-images of each image which are then multiplexed, or interleaved, in frames of a video stream to be encoded and then decoded using a standardized video encoder. A system for viewing may utilize a standard video decoder and a formatting device that de-interleaves the decoded sub-images of each frame reformats the images for a display device. A 3D video may be encoded using a most advantageous interleaving format such that a preferred quality and compression ratio is reached. In one embodiment, the invention includes a display device that accepts data in multiple formats.

Подробнее
01-05-2013 дата публикации

Method and system for illumination compensation and transition for video coding and processing

Номер: CN103081468A
Принадлежит:

Deriving illumination compensation parameters and detection of illumination dominant transitions types for video coding and processing applications is described. Illumination changes such as fade-ins, fade-outs, cross-fades, and flashes are detected. Detection of these illumination changes is then used for weighted prediction to provide for improved illumination compensation.

Подробнее
03-10-2012 дата публикации

High performance rate control for multi-layered video coding applications

Номер: CN102714725A
Принадлежит:

Methods and systems for multi-layer video encoding are described. In the system, an inter-layer analysis module gathers and analyzes inter-layer coding statistics, so that a rate control module can perform rate control on the various layers based on the statistics.

Подробнее
10-12-2013 дата публикации

Reconstruction of de-interleaved views, using adaptive interpolation based on disparity between the views for up-sampling

Номер: US0008605782B2

Method for reconstructing interleaved views of a 3D image comprising the steps of: receiving interleaved views of a 3D image, de-interleaving said interleaved views into a first view and a second view, each de-interleaved view comprising missing pixels, at each pixel location determining an amount of disparity between said first and second views, for each view interpolating each missing pixel based on the amount of disparity determined at the associated pixel location, wherein when the determined disparity is above a predetermined disparity threshold, the missing pixel value is interpolated from the data of the view to be interpolated only (i.e. the view comprising said missing pixel value), and when the determined disparity is at or below said predetermined disparity threshold, the missing pixel value is interpolated from the data of both views.

Подробнее
02-02-2017 дата публикации

VIDEO COMPRESSION AND TRANSMISSION TECHNIQUES

Номер: US20170034513A1
Принадлежит:

Embodiments feature families of rate allocation and rate control methods that utilize advanced processing of past and future frame/field picture statistics and are designed to operate with one or more coding passes. At least two method families include: a family of methods for a rate allocation with picture look-ahead; and a family of methods for average bit rate (ABR) control methods. At least two other methods for each method family are described. For the first family of methods, some methods may involve intra rate control. For the second family of methods, some methods may involve high complexity ABR control and/or low complexity ABR control. These and other embodiments can involve any of the following: spatial coding parameter adaptation, coding prediction, complexity processing, complexity estimation, complexity filtering, bit rate considerations, quality considerations, coding parameter allocation, and/or hierarchical prediction structures, among others. 1. (canceled)2. A method of coding a picture for video , the method comprising:estimating, with a pre-analyzer, a complexity of a current picture based upon analyzing a complexity of past pictures or future pictures;determining, with the pre-analyzer, a level of correlation for the current picture with the past pictures or the future pictures; andbased upon the determination of the level of correlation, coding the current picture at a quality level that is higher than a quality level of the past pictures or the future pictures.3. The method of claim 2 , further comprising coding claim 2 , with a coder having motion-compensated prediction claim 2 , the past pictures or future pictures at the quality level of the current picture claim 2 , wherein the level of correlation comprises a level of temporal correlation.4. The method of claim 2 , further comprising determining claim 2 , with the pre-analyzer claim 2 , a metric that indicates temporal correlation for the current picture with the past pictures or the ...

Подробнее
03-10-2012 дата публикации

Multiple-pass rate control for video coding applications

Номер: CN102714724A
Принадлежит:

Methods and systems for coding applications are described, where multiple coding passes are performed for purposes of rate control. Prediction structures are generated and used by the multiple coding passes to update coding statistics generated prior to the coding passes. The updated coding statistics are used for a subsequent coding operation.

Подробнее
14-02-2017 дата публикации

Spatiotemporal prediction for bidirectionally predictive (B) pictures and motion vector prediction for multi-picture reference motion compensation

Номер: US9571854B2

Several improvements for use with Bidirectionally Predictive (B) pictures within a video sequence are provided. In certain improvements Direct Mode encoding and/or Motion Vector Prediction are enhanced using spatial prediction techniques. In other improvements Motion Vector prediction includes temporal distance and subblock information, for example, for more accurate prediction. Such improvements and other presented herein significantly improve the performance of any applicable video coding system/logic.

Подробнее
22-08-2012 дата публикации

Speedup techniques for rate distortion optimized quantization

Номер: CN102648627A
Принадлежит:

Techniques for selecting a coding mode for an image coding process are described. Coding modes can be selected through a coding mode transition state machine, a re-quantization process, selection of an optimal transform size, by skipping some quantization parameters, or by performing motion search.

Подробнее
23-01-2018 дата публикации

Coding and decoding of interleaved image data

Номер: US0009877046B2

Sampled data is packaged in checkerboard format for encoding and decoding. The sampled data may be quincunx sampled multi-image video data (e.g., 3D video or a multi-program stream), and the data may also be divided into sub-images of each image which are then multiplexed, or interleaved, in frames of a video stream to be encoded and then decoded using a standardized video encoder. A system for viewing may utilize a standard video decoder and a formatting device that de-interleaves the decoded sub-images of each frame reformats the images for a display device. A 3D video may be encoded using a most advantageous interleaving format such that a preferred quality and compression ratio is reached. In one embodiment, the invention includes a display device that accepts data in multiple formats.

Подробнее
30-05-2017 дата публикации

Predictive motion vector coding

Номер: US0009667993B2

Overlapped block disparity estimation and compensation is described. Compensating for images with overlapped block disparity compensation (OBDC) involves determining if OBDC is enabled in a video bit stream, and determining if OBDC is enabled for one or more macroblocks that neighbor a first macroblock within the video bit stream. The neighboring macroblocks may be transform coded. If OBDC is enabled in the video bit stream and for the one or more neighboring macroblocks, predictions may be made for a region of the first macroblock that has an edge adjacent with the neighboring macroblocks. OBDC can be causally applied. Disparity compensation parameters or modes may be shared amongst views or layers. A variety of predictions may be used with causally-applied OBDC.

Подробнее
21-08-2013 дата публикации

Enhancement methods for sampled and multiplexed image and video data

Номер: CN103262532A
Принадлежит:

Enhancement methods for sampled and multiplexed image and video data are described. Each component picture is separately processed either after de-multiplexing or on the fly. Processing and de-multiplexing can be combined in a single joint step. The methods apply to both encoding and decoding system and include applications to scalable video coding systems.

Подробнее
08-08-2017 дата публикации

Speedup techniques for rate distortion optimized quantization

Номер: US0009729888B2

Techniques for selecting a coding mode for an image coding process are described. Coding modes can be selected through a coding mode transition state machine, a re-quantization process, selection of an optimal transform size, by skipping some quantization parameters, or by performing motion search.

Подробнее
06-09-2016 дата публикации

Enhancement methods for sampled and multiplexed image and video data

Номер: US0009438881B2

Enhancement methods for sampled and multiplexed image and video data are described. Each component picture is separately processed either after de-multiplexing or on the fly. Processing and de-multiplexing can be combined in a single joint step. The methods apply to both encoding and decoding system and include applications to scalable video coding systems.

Подробнее
20-10-2011 дата публикации

Reconstruction of De-Interleaved Views, Using Adaptive Interpolation Based on Disparity Between the Views for Up-Sampling

Номер: US20110254921A1

Method for reconstructing interleaved views of a 3D image comprising the steps of: receiving interleaved views of a 3D image, de-interleaving said interleaved views into a first view and a second view, each de-interleaved view comprising missing pixels, at each pixel location determining an amount of disparity between said first and second views, for each view interpolating each missing pixel based on the amount of disparity determined at the associated pixel location, wherein when the determined disparity is above a predetermined disparity threshold, the missing pixel value is interpolated from the data of the view to be interpolated only (i.e. the view comprising said missing pixel value), and when the determined disparity is at or below said predetermined disparity threshold, the missing pixel value is interpolated from the data of both views.

Подробнее
17-01-2017 дата публикации

Context based inverse mapping method for layered codec

Номер: US0009549194B2

Context based inverse mapping methods are provided. An image with pixel values within a lower dynamic range can be mapped to an image with pixel values within a higher dynamic range by utilizing context information associated with pixels in the lower dynamic range image.

Подробнее
13-09-2016 дата публикации

Video compression and transmission techniques

Номер: US0009445110B2

Embodiments feature families of rate allocation and rate control methods that utilize advanced processing of past and future frame/field picture statistics and are designed to operate with one or more coding passes. At least two method families include: a family of methods for a rate allocation with picture look-ahead; and a family of methods for average bit rate (ABR) control methods. At least two other methods for each method family are described. For the first family of methods, some methods may involve intra rate control. For the second family of methods, some methods may involve high complexity ABR control and/or low complexity ABR control. These and other embodiments can involve any of the following: spatial coding parameter adaptation, coding prediction, complexity processing, complexity estimation, complexity filtering, bit rate considerations, quality considerations, coding parameter allocation, and/or hierarchical prediction structures, among others.

Подробнее
15-09-2016 дата публикации

PREDICTIVE MOTION VECTOR CODING

Номер: US20160269743A1
Принадлежит:

Overlapped block disparity estimation and compensation is described. Compensating for images with overlapped block disparity compensation (OBDC) involves determining if OBDC is enabled in a video bit stream, and determining if OBDC is enabled for one or more macroblocks that neighbor a first macroblock within the video bit stream. The neighboring macroblocks may be transform coded. If OBDC is enabled in the video bit stream and for the one or more neighboring macroblocks, predictions may be made for a region of the first macroblock that has an edge adjacent with the neighboring macroblocks. OBDC can be causally applied. Disparity compensation parameters or modes may be shared amongst views or layers. A variety of predictions may be used with causally-applied OBDC.

Подробнее
24-01-2017 дата публикации

Method and system for illumination compensation and transition for video coding and processing

Номер: US0009554156B2

Deriving illumination compensation parameters and detection of illumination dominant transitions types for video coding and processing applications is described. Illumination changes such as fade-ins, fade-outs, cross-fades, and flashes are detected. Detection of these illumination changes is then used for weighted prediction to provide for improved illumination compensation.

Подробнее
06-02-2018 дата публикации

Video coding

Номер: US0009888237B2

Improved video coding is described to encode video data within a sequence of video frames. To this end, at least a portion of a reference frame is encoded to include motion information associated with the portion of the reference frame. At least a portion of a predictable frame that includes video data predictively correlated to s portion of said reference frame is defined based on the motion information. At least said portion of the predictable frame is encoded without including corresponding motion information and including mode identifying data. The mode identifying data indicate that the encoded portion of the predictable frame can be directly derived using at least the motion information associated with the portion of the reference frame.

Подробнее
21-09-2017 дата публикации

DEVICE AND METHOD FOR FAST BLOCK-MATCHING MOTION ESTIMATION IN VIDEO ENCODERS

Номер: US20170272776A1
Принадлежит:

A solution is provided to estimate motion vectors of a video. A multistage motion vector prediction engine is configured to estimate multiple best block-matching motion vectors for each block in each video frame of the video. For each stage of the motion vector estimation for a block of a video frame, the prediction engine selects a test vector form a predictor set of test vectors, computes a rate-distortion optimization (RDO) based metric for the selected test vector, and selects a subset of test vectors as individual best matched motion vectors based on the RDO based metric. The selected individual best matched motion vectors are compared and a total best matched motion vector is selected based on the comparison. The prediction engine selects iteratively applies one or more global matching criteria to the selected best matched motion vector to select a best matched motion vector for the block of pixels.

Подробнее
11-10-2016 дата публикации

Complexity-adaptive scalable decoding and streaming for multi-layered video systems

Номер: US0009467690B2

Adaptive decoding and streaming multi-layer video systems and methods are described. The decoding systems comprise a base layer decoder and one or more enhancement layer decoders. The streaming systems comprise a base layer packetizer and one or more enhancement layer packetizers. A decoding adaptor controls operation of the base layer and/or enhancement layer decoders. A packetizing adaptor controls operation of the base layer and/or enhancement layer packetizers.

Подробнее
12-12-2017 дата публикации

Predictive motion vector coding

Номер: US0009843807B2

Overlapped block disparity estimation and compensation is described. Compensating for images with overlapped block disparity compensation (OBDC) involves determining if OBDC is enabled in a video bit stream, and determining if OBDC is enabled for one or more macroblocks that neighbor a first macroblock within the video bit stream. The neighboring macroblocks may be transform coded. If OBDC is enabled in the video bit stream and for the one or more neighboring macroblocks, predictions may be made for a region of the first macroblock that has an edge adjacent with the neighboring macroblocks. OBDC can be causally applied. Disparity compensation parameters or modes may be shared amongst views or layers. A variety of predictions may be used with causally-applied OBDC.

Подробнее
05-10-2017 дата публикации

COMPATIBLE STEREOSCOPIC VIDEO DELIVERY

Номер: US20170289520A1
Принадлежит:

Stereoscopic images are subsampled and placed in a “checkerboard” pattern in an image. The image is encoded in a monoscopic video format. The monoscopic video is transmitted to a device where the “checkerboard” is decoded. Portions of the checkerboard (e.g., “black” portions) are used to reconstruct one of the stereoscopic images and the other portion of the checkerboard (e.g., “white” portions) are used to reconstruct the other image. The subsamples are, for example, taken from the image in a location coincident to the checkerboard position in which the subsamples are encoded.

Подробнее
29-11-2016 дата публикации

Chroma quantization in video coding

Номер: US0009510002B2
Принадлежит: APPLE INC., APPLE INC, Apple Inc.

A method of signaling additional chroma QP offset values that are specific to quantization groups is provided, in which each quantization group explicitly specifies its own set of chroma QP offset values. Alternatively, a table of possible sets of chroma QP offset values is specified in the header area of the picture, and each quantization group uses an index to select an entry from the table for determining its own set of chroma QP offset values. The quantization group specific chroma QP offset values are then used to determine the chroma QP values for blocks within the quantization group in addition to chroma QP offset values already specified for higher levels of the video coding hierarchy.

Подробнее
26-12-2017 дата публикации

Video encoding optimization with extended spaces

Номер: US0009854246B2
Принадлежит: APPLE INC., APPLE INC, Apple Inc.

Embodiments of the present invention may provide a video coder. The video coder may include an encoder to perform coding operations on a video signal in a first format to generate coded video data, and a decoder to decode the coded video data. The video coder may also include an inverse format converter to convert the decoded video data to second format that is different than the first format and an estimator to generate a distortion metric using the decoded video data in the second format and the video signal in the second format. The encoder may adjust the coding operations based on the distortion metric.

Подробнее
03-04-2013 дата публикации

Reference processing using advanced motion models for video coding

Номер: CN103026707A
Принадлежит:

Processing a reference picture is described. A reference processing unit enables signaling of parameters such as motion model parameters, interpolation filter parameters, intensity compensation parameters, and denoising filter parameters. Methods for estimating the various parameters are also discussed. Processing improves quality of a reference picture prior to its use for prediction of a subsequent picture and thus improves the prediction.

Подробнее
11-05-2017 дата публикации

ADAPTIVE CHROMA DOWNSAMPLING AND COLOR SPACE CONVERSION TECHNIQUES

Номер: US20170134731A1
Принадлежит:

Techniques for selecting a luminance value for color space conversion are disclosed. Techniques include determining values for Cb and Cr from values for R′, G′, and B′; producing a reconstructed Cb* value and a reconstructed Cr* value by processing the Cb and Cr values; and determining a plurality of Y′ value options from the values for Cb* and Cr*. A Y′ output value may be selected based on the plurality of Y′ value options.

Подробнее
03-04-2013 дата публикации

Systems and methods for multi-layered frame-compatible video delivery

Номер: CN103026706A
Принадлежит:

Multi-layered frame-compatible video delivery is described. Multi-layered encoding and decoding methods, comprising a base layer and at least one enhancement layer with reference processing, are provided. In addition, multi-layered encoding and decoding methods with inter-layer dependencies are described. Encoding and decoding methods that are capable of frame-compatible 3D video delivery are also described.

Подробнее
03-01-2017 дата публикации

Pre-processing for bitdepth and color format scalable video coding

Номер: US0009538176B2

Methods for scalable video coding are described. Such methods can be used to deliver video contents in Low Dynamic Range (LDR) and/or one color format and then converting the video contents to High Dynamic Range (HDR) and/or a different color format, respectively, while pre-processing video content.

Подробнее
18-07-2012 дата публикации

Adaptive interpolation filters for multi-layered video delivery

Номер: CN102598660A
Принадлежит:

Adaptive interpolation filters for multi-layered video delivery are described. Combinations of such adaptive filters with other pre-processing steps in the context of frame or 2D compatible 3D and 2D scalable video delivery system are also addressed.

Подробнее
12-06-2013 дата публикации

Joint layer optimization for a frame-compatible video delivery

Номер: CN103155559A
Принадлежит:

Joint layer optimization for a frame-compatible video delivery is described. More specifically, methods for efficient mode decision, motion estimation, and generic encoding parameter selection in multiple-layer codecs that adopt a reference processing unit (RPU) to exploit inter-layer correlation to improve coding efficiency as described.

Подробнее
13-12-2016 дата публикации

Optimized filter selection for reference picture processing

Номер: US0009521413B2

Reference processing may be used in a video encoder or decoder to derive reference pictures that are better correlated with a source image to be encoded or decoded, which generally yields better coding efficiency. Methods for filter selection for a reference processing unit adapted for use in a video codec system are discussed. Specifically, methods for filter selection based on performing motion estimation and obtaining distortion/cost information by comparing reference pictures, either processed or non-processed, with the source image to be encoded are discussed.

Подробнее
19-06-2013 дата публикации

Error resilient rate distortion optimization for image and video encoding

Номер: CN103168469A
Принадлежит:

Error resilient rate distortion optimization (ERRDO) is used for transmitting high quality images and video over constrained bandwidth networks, e.g., in streaming applications. Transmitting high quality images and video by reducing computational complexity is described.

Подробнее
26-09-2017 дата публикации

Encoding and decoding architectures for format compatible 3D video delivery

Номер: US0009774882B2

Encoding and decoding architectures for 3D video delivery are described, such as 2D compatible 3D video delivery and frame compatible 3D video delivery. The architectures include pre-processing stages to pre-process the output of a base layer video encoder and/or decoder and input the pre-processed output into an enhancement layer video encoder and/or decoder of one or more enhancement layers. Multiplexing methods of how to combine the base and enhancement layer videos are also described.

Подробнее
10-08-2017 дата публикации

PREDICTIVE MOTION VECTOR CODING

Номер: US20170230666A1
Принадлежит:

Overlapped block disparity estimation and compensation is described. Compensating for images with overlapped block disparity compensation (OBDC) involves determining if OBDC is enabled in a video bit stream, and determining if OBDC is enabled for one or more macroblocks that neighbor a first macroblock within the video bit stream. The neighboring macroblocks may be transform coded. If OBDC is enabled in the video bit stream and for the one or more neighboring macroblocks, predictions may be made for a region of the first macroblock that has an edge adjacent with the neighboring macroblocks. OBDC can be causally applied. Disparity compensation parameters or modes may be shared amongst views or layers. A variety of predictions may be used with causally-applied OBDC.

Подробнее
29-05-2013 дата публикации

Treating video information

Номер: CN103124354A
Принадлежит:

The invention discloses treating video information. Systems, methods, and techniques for treating video information are described. In one implementation, a method includes receiving video information, classifying one or more frames in the received video information as a scene, adjusting one or more coding parameters based on the classification of the frames, and coding the video information in accordance with the adjusted coding parameters.

Подробнее
11-10-2016 дата публикации

Multiple-pass rate control for video coding applications

Номер: US0009467693B2

Methods and systems for coding applications are described, where multiple coding passes are performed for purposes of rate control. Prediction structures are generated and used by the multiple coding passes to update coding statistics generated prior to the coding passes. The updated coding statistics are used for a subsequent coding operation.

Подробнее
10-10-2012 дата публикации

Methods and systems for reference processing in image and video codecs

Номер: CN102726046A
Принадлежит:

Multi-layer encoding and decoding systems and methods are provided. A processing module processes outputs of a first base or enhancement layer and sends the processed outputs to a second, enhancement layer. Operation of the processing module is controlled, so that the second layer can receive processed or unprocessed outputs of the first layer in accordance with the circumstances. Processing of the outputs of the first layer can occur together with or separately from a disparity compensation process.

Подробнее
25-10-2016 дата публикации

Systems and methods for multi-layered frame-compatible video delivery

Номер: US0009479772B2

Multi-layered frame-compatible video delivery is described. Multi-layered encoding and decoding methods, comprising a base layer and at least one enhancement layer with reference processing, are provided. In addition, multi-layered encoding and decoding methods with inter-layer dependencies are described. Encoding and decoding methods that are capable of frame-compatible 3D video delivery are also described.

Подробнее
04-07-2017 дата публикации

Rate control for video splicing applications

Номер: US0009699454B2

Rate control techniques are provided for encoding an input video sequence into a compressed coded bitstream with multiple coding passes. The final coding pass may comprise final splices with non-overlapping frames that do not extend into neighboring final splices. A final splice in the final coding pass may correspond to at least one non-final splice in a non-final coding pass. A non-final splice may have overlapping frames that extend into neighboring final splices in the final coding pass. The overlapping frames in the non-final splice may be used to derive complexity information about the neighboring final splices. The complexity information about the neighboring final splices, as derived from the overlapping frames, may be used to allocate or improve rate control related budgets in encoding the final splice into the compressed coded bitstream in the final coding pass.

Подробнее
21-03-2017 дата публикации

Method and apparatus for adaptive group of pictures (GOP) structure selection

Номер: US0009602840B2

There are provided a method and apparatus for adaptive Group of Pictures structure selection. The apparatus includes an encoder for encoding a video sequence using a Group of Pictures structure by performing, for each Group of Pictures for the video sequence, picture coding order selection, picture type selection, and reference picture selection. The selections are based upon a Group of Pictures length.

Подробнее
30-01-2013 дата публикации

Video compression under multiple distortion constraints

Номер: CN102905131A
Принадлежит:

Video compression is performed under multiple distortion constraints. Video coding includes determining a first set of Lagrangian cost values for multiple coding modes using a first distortion metric, determining a second set of Lagrangian cost values for the multiple coding modes using a second distortion metric, and selecting one of the coding modes based on the first set of Lagrangian cost values and the second set of Lagrangian cost values to encode a pixel block using the selected coding mode. The distortion metrics can include information associated with video display characteristics such as screen display size, video processing performance, a distortion characteristic, a temporal characteristic, or a spatial characteristic. The distortion metrics can also include characteristics of multiple video displays, such as parameters for usage, importance, design, and the technology type of the video displays.

Подробнее
12-12-2017 дата публикации

Compatible stereoscopic video delivery

Номер: US0009843785B2

Stereoscopic images are subsampled and placed in a “checkerboard” pattern in an image. The image is encoded in a monoscopic video format. The monoscopic video is transmitted to a device where the “checkerboard” is decoded. Portions of the checkerboard (e.g., “black” portions) are used to reconstruct one of the stereoscopic images and the other portion of the checkerboard (e.g., “white” portions) are used to reconstruct the other image. The subsamples are, for example, taken from the image in a location coincident to the checkerboard position in which the subsamples are encoded.

Подробнее
06-03-2018 дата публикации

Compatible stereoscopic video delivery

Номер: US0009912931B1

Stereoscopic images are subsampled and placed in a “checkerboard” pattern in an image. The image is encoded in a monoscopic video format. The monoscopic video is transmitted to a device where the “checkerboard” is decoded. Portions of the checkerboard (e.g., “black” portions) are used to reconstruct one of the stereoscopic images and the other portion of the checkerboard (e.g., “white” portions) are used to reconstruct the other image. The subsamples are, for example, taken from the image in a location coincident to the checkerboard position in which the subsamples are encoded.

Подробнее
19-09-2017 дата публикации

Adaptive color space transform coding

Номер: US9769480B2
Принадлежит: APPLE INC, Apple Inc.

An encoder system may include an analyzer that analyzes a current image area in an input video to select a transform. A selectable residue transformer, controlled by the analyzer, may perform the selectable transform on a residue image generated from the current image area and a predicted current image area, to generate a transformed residue image. An encoder may encode the transformed residue image to generate output data. The analyzer controls the encoder to encode information to identify the selectable transform and to indicate that the selectable transform for the current image area is different from a transform of a previous image area of the input video. A decoder system may include components appropriate for decoding the output data from the encoder system.

Подробнее
06-06-2017 дата публикации

Device and method for fast block-matching motion estimation in video encoders

Номер: US0009674548B2
Принадлежит: FastVDO LLC, FASTVDO LLC

A solution is provided to estimate motion vectors of a video. A multistage motion vector prediction engine is configured to estimate multiple best block-matching motion vectors for each block in each video frame of the video. For each stage of the motion vector estimation for a block of a video frame, the prediction engine selects a test vector form a predictor set of test vectors, computes a rate-distortion optimization (RDO) based metric for the selected test vector, and selects a subset of test vectors as individual best matched motion vectors based on the RDO based metric. The selected individual best matched motion vectors are compared and a total best matched motion vector is selected based on the comparison. The prediction engine selects iteratively applies one or more global matching criteria to the selected best matched motion vector to select a best matched motion vector for the block of pixels.

Подробнее
02-01-2013 дата публикации

Complexity-adaptive scalable decoding and streaming for multi-layered video systems

Номер: CN102860008A
Принадлежит:

Adaptive decoding and streaming multi-layer video systems and methods are described. The decoding systems comprise a base layer decoder and one or more enhancement layer decoders. The streaming systems comprise a base layer packetizer and one or more enhancement layer packetizers. A decoding adaptor controls operation of the base layer and/or enhancement layer decoders. A packetizing adaptor controls operation of the base layer and/or enhancement layer packetizers.

Подробнее
02-02-2012 дата публикации

Directed Interpolation and Data Post-Processing

Номер: US20120026288A1

An encoding device evaluates a plurality of processing and/or post-processing algorithms and/or methods to be applied to a video stream, and signals a selected method, algorithm, class or category of methods/algorithms either in an encoded bit-stream or as side information related to the encoded bitstream. A decoding device or post-processor utilizes the signaled algorithm or selects an algorithm/method based on the signaled method or algorithm. The selection is based, for example, on availability of the algorithm/method at the decoder/post-processor and/or cost of implementation. The video stream may comprise, for example, downsampled multiplexed stereoscopic images and the selected algorithm may include any of upconversion and/or error correction techniques that contribute to a restoration of the downsampled images. 1. A method comprising the step of receiving a signal indicative of a processing technique to be utilized on a data stream comprising a video comprising at least two views.2. The method according to claim 1 , wherein the processing technique comprises resolution conversion comprising at least one of spatial resolution conversion and temporal resolution conversion.3. The method according to claim 1 , wherein the processing technique comprises scaling comprising at least one of de-interlacing claim 1 , temporal interpolation claim 1 , and spatial interpolation.4. The method according to claim 1 , wherein the processing technique comprises error concealment comprising a prediction of motion of a region in error.5. The method according to claim 1 , wherein the signal is further indicative of at least one of a spatial and temporal extent to which the processing technique is to be applied claim 1 , and the data stream is a video data stream and the extent comprises whether the processing technique is applied to at least one of all of the video claim 1 , portions of the video claim 1 , pictures of the video claim 1 , and regions within the video.6. The method ...

Подробнее
02-02-2012 дата публикации

Adaptive Interpolation Filters for Multi-Layered Video Delivery

Номер: US20120027079A1
Принадлежит: Dolby Laboratories Licensing Corp

Adaptive interpolation filters for multi-layered video delivery are described. Combinations of such adaptive filters with other pre-processing steps in the context of frame or 2D compatible 3D and 2D scalable video delivery system are also addressed.

Подробнее
09-02-2012 дата публикации

Filter Selection for Video Pre-Processing in Video Applications

Номер: US20120033040A1

Filter selection methods and filter selectors for video pre-processing in video applications are described. A region of an input image is pre-processed by multiple pre-processing filters and the selection of the pre-processing filter for subsequent coding is based on the evaluated metric of the region. 1. A method for selecting a downsampling filter for scalable video delivery , comprising:inputting one or more input images into a plurality of downsampling filters to form, for each downsampling filter, an output image or data stream;encoding the output image or data stream to form an encoded and reconstructed image or data stream, wherein the encoding comprises a base layer encoding and an enhancement layer encoding;for each downsampling filter, evaluating a metric of the encoded and reconstructed image or data stream; andselecting a downsampling filter among the plurality of downsampling filters based on the evaluated metric for each downsampling filter and feedback from the enhancement layer encoding, wherein the feedback comprises information on adaptive upsampling filter parameters used for base layer to enhancement layer prediction.2. The method as recited in claim 1 , further comprising 3D-interleaving the output of each downsampling filter.3. The method as recited in claim 1 , wherein selection of a downsampling filter based on the evaluated metric for each filter allows selection of the output encoded data stream corresponding to the selected downsampling filter.4. The method as recited in claim 1 , further comprising performing a two-stage encoding claim 1 , a first stage encoding occurring before selecting the downsampling filter and a second stage encoding occurring after selecting the downsampling filter.5. The method as recited in claim 1 , wherein scalable video is delivered with one or more of different bit-depths claim 1 , scales claim 1 , or color space representations.6. The method as recited in claim 1 , wherein the first stage encoding comprises ...

Подробнее
19-04-2012 дата публикации

METHOD FOR EMBEDDING SUBTITLES AND/OR GRAPHIC OVERLAYS IN A 3D OR MULTI-VIEW VIDEO DATA

Номер: US20120092454A1
Принадлежит:

A method for embedding subtitles and/or graphic overlays in a 3D or multi-view video application is described. The subtitles and/or graphic overlays are provided separately for each view of the 3D or multi-view video application. The views with the subtitles and/or graphic overlays are then processed to form a subtitled and/or graphic overlaid 3D or multi-view video application. 1. A method for embedding subtitles and/or graphic overlays in a 3D or multi-view video application to form a subtitled and/or overlaid 3D or multi-view video application , comprising:providing a plurality of two or more views, wherein view disparities are present between the views in order to allow subsequent formation of the 3D or multi-video application;providing the subtitles and/or graphic overlays separately for each view based on the view disparities, wherein at least one view provided with subtitles and/or graphic overlays with a different representation from the subtitle and/or graphic overlays of at least one other view among the plurality of two or more views; andprocessing the plurality of two or more views and the subtitles and/or graphic overlays to form the subtitled and/or overlaid 3D or multi-view video applicationwherein the different representation comprises at least one of different color, sharpness, and outlines.2. A method for embedding subtitles and/or graphic overlays in a 3D or multi-view video application , comprising:providing a plurality of two or more views, wherein view disparities are present between the views in order to allow subsequent formation of the 3D or multi-video application;providing the subtitles and/or graphic overlays separately for each view based on the view disparities;converting the provided subtitles and/or graphic overlays as metadata; andprocessing the plurality of two or more views and the subtitles and/or graphic overlays based on the metadata to form a subtitled and/or overlaid 3D or multi-view video application,wherein the metadata ...

Подробнее
12-07-2012 дата публикации

Speedup Techniques for Rate Distortion Optimized Quantization

Номер: US20120177109A1

Techniques for selecting a coding mode for an image coding process are described. Coding modes can be selected through a coding mode transition state machine, a re-quantization process, selection of an optimal transform size, by skipping some quantization parameters, or by performing motion search. 1. A method of selecting a coding mode and a quantization parameter for an image coding unit to be coded and quantized in an image coding process , comprising:selecting an initial quantization parameter and a best coding mode for the initial quantization parameter;evaluating coding modes for subsequent quantization parameters in accordance with a coding mode transition state machine comprising coding mode states and transitions between coding mode states, wherein each state of the coding mode transition state machine corresponds to a coding mode and a mode transition likelihood score is assigned to the transition between a pair of coding mode states; andselecting an optimal coding mode and an optimal quantization parameter among the evaluated coding modes and quantization parameters.2. The method as recited in claim 1 , wherein the mode transition likelihood score is higher if a transition between coding mode states is more likely and is lower if the transition between coding mode states is less likely.3. The method as recited in claim 1 , wherein the mode transition likelihood score is assigned in accordance with characteristics of the image coding unit.4. The method as recited in claim 3 , wherein the characteristic of the image is analyzed and determined by image pre-analysis prior to image coding.5. The method as recited in claim 1 , wherein the mode transition likelihood score is assigned in accordance with the type of transition between coding modes.6. The method as recited in claim 1 , wherein the number of transitions between coding mode states is controllable.7. The method as recited in claim 6 , wherein the number of transitions between coding mode states is ...

Подробнее
08-11-2012 дата публикации

Multimedia Coding and Decoding with Additional Information Capability

Номер: US20120281751A1

A multimedia coding and decoding system and method is presented that uses the specific prediction mode to signal supplemental information, e.g., metadata, while considering and providing trade offs between coding performance and metadata capacity. The prediction mode can be encoded according to a mode table that relates mode to bits and by considering coding impact. Start and stop codes can be used to signal the message, while various techniques of how to properly design the mode to bits tables are presented. 1. A computer-implemented method for encoding a media signal , comprising:receiving a media signal;obtaining supplemental information to be encoded within said media signal;using said supplemental information to select one encoding type from a plurality of different encoding types; andencoding said media signal using said one encoding type, where the encoding type represents the supplemental information.2. A method as in claim 1 , wherein said media signal is a video signal.3. A method as in claim 2 , further comprising defining at least one of a start code claim 2 , an end code claim 2 , or a length code claim 2 , and using said encoding type to represent said at least one of said start code claim 2 , end code claim 2 , or length code within the video signal location adjacent the supplemental information.4. A method as in claim 3 , wherein said start code or end code represent sequences of encoding decisions which are unlikely to occur in real video.5. A method as in claim 2 , wherein said supplemental information is related to contents of the video signal claim 2 , and is temporally synchronized with different portions of the video signal.6. A method as in claim 2 , further comprising detecting a first encoding type that is selected based on the secondary information claim 2 , in which the first encoding type causes degradation in the video claim 2 , and overriding said selecting based on said detecting.7. A method as in claim 6 , wherein said overriding said ...

Подробнее
22-11-2012 дата публикации

Filtering for Image and Video Enhancement Using Asymmetric Samples

Номер: US20120293620A1

Methods and devices for filtering lower quality images or sequences of images with higher quality images or sequences of images are described. 156-. (canceled)57. A method for processing samples of an image or sequence of images , the method comprising:encoding a first set of samples with a first encoding process;encoding a second set of samples with a second encoding process;filtering reconstructed samples of the first set based on information associated with at least one of the first encoding process and the second encoding process; andcombining the filtered reconstructed samples of the first set with reconstructed samples of the second set to obtain a new representation of the filtered reconstructed samples of the first set.58. The method as recited in claim 57 , wherein the first set of samples and the second set of samples are taken from different layers of a scalable video delivery application.59. The method as recited in claim 58 , wherein one set of samples is taken from a base layer bitstream and the other set of samples is taken from an enhancement layer.60. The method as recited in claim 57 , wherein the first set of samples and the second set of samples are taken from different fields of an interlaced image sequence or different views of an interleaved 3D or multiview image sequence.61. The method as recited in claim 60 , wherein the samples of the first set correspond to even fields of the interlaced or interleaved image sequence and the samples of the second correspond to odd fields of the interlaced or interleaved image sequence or vice versa.62. The method as recited in claim 60 , wherein the interlaced or interleaved image sequence comprises a 3D image sequence claim 60 , the samples of the first set correspond to left stereoscopic views of the interlaced or interleaved image sequence and the samples of the second set correspond to right stereoscopic views of the interlaced or interleaved image sequence or vice versa.63. The method as recited in ...

Подробнее
03-01-2013 дата публикации

METHOD, SYSTEM AND DEVICE FOR IMPROVING VIDEO QUALITY THROUGH IN-LOOP TEMPORAL PRE-FILTERING

Номер: US20130003872A1
Принадлежит:

A method, system and device are provided for pre-filtering device for filtering a video signal prior to digitally encoding. The method includes receiving at least one input picture and at least one reconstructed picture from an encoding process and performing an in-loop temporal filtering process using at least one input picture and at least one reconstructed picture from an encoding process to output a pre-filtered video signal for use in an encoding process. The result is enabling an encoding process to produce an output with increased temporal correlation between adjacent pictures regardless of the coding type, since the artifacts introduced by the encoding process are also considered by the process. 1. (canceled)2. (canceled)3. (canceled)4. (canceled)5. (canceled)6. (canceled)7. (canceled)8. (canceled)9. (canceled)10. (canceled)11. A method for a pre-filtering device to filter a video signal prior to digitally encoding , comprising:receiving at least one input picture;receiving at least one reconstructed picture from an encoding process; andperforming an in-loop temporal filtering process using at least one input picture and at least one reconstructed picture from an encoding process to output a pre-filtered video signal for use in an encoding process, wherein picture statistics are applied in the in-loop temporal filtering process.12. The method according to claim 1 , wherein the in-loop temporal filtering process includes generating the picture statistics.13. The method according to claim 1 , further comprising receiving picture statistics from an external source claim 1 , wherein the in-loop temporal filtering process includes separately generating the picture statistics.14. The method according to claim 1 , further comprising receiving picture statistics from a video pre-process.15. The method according to claim 1 , further comprising receiving the picture statistics from a video pre-filtering process.16. The method according to claim 1 , further comprising ...

Подробнее
31-01-2013 дата публикации

High Performance Rate Control for Multi-Layered Video Coding Applications

Номер: US20130028316A1

Methods and systems for multi-layer video encoding are described. In the system, an inter-layer analysis module gathers and analyzes inter-layer coding statistics, so that a rate control module can perform rate control on the various layers based on the statistics. 129-. (canceled)30. A multi-layer video encoding system , comprising:a base layer encoder to provide a base layer encoded bitstream of a video input;one or more enhancement layer encoders to provide corresponding one or more enhancement layer encoded bitstreams of the video input;an inter-layer analysis module to gather and analyze coding statistics from one or more of the base layer encoder and one or more enhancement layer encoders, provide one or more rate control models, and update the one or more provided rate control models based on the coding statistics, each rate control model being associated with the base layer encoder and/or the one or more enhancement layer encoders; andat least one rate control module to perform rate control on one or more of the base layer encoder and one or more enhancement layer encoders based on the inter-layer analysis module, the at least one rate control module to provide one or more rate control models, each rate control model being associated with base layer encoding by the base layer encoder and/or enhancement layer encoding by the one or more enhancement layer encoders;wherein:rate control models among the one or more rate control models associated with the enhancement layer encoding are updated during the base layer encoding using inter-layer relationships; andeach of the one or more rate control models is updated during the enhancement layer encoding using inter-layer relationships.31. The system as recited in claim 30 , further comprising a preanalysis module claim 30 , wherein the coding statistics analyzed by the inter-layer analysis module also depends on the preanalysis module.32. The system as recited in claim 30 , wherein rate control of a layer is based ...

Подробнее
25-04-2013 дата публикации

Complexity-Adaptive Scalable Decoding and Streaming for Multi-Layered Video Systems

Номер: US20130101015A1

Adaptive decoding and streaming multi-layer video systems and methods are described. The decoding systems comprise a base layer decoder and one or more enhancement layer decoders. The streaming systems comprise a base layer packetizer and one or more enhancement layer packetizers. A decoding adaptor controls operation of the base layer and/or enhancement layer decoders. A packetizing adaptor controls operation of the base layer and/or enhancement layer packetizers. 135-. (canceled)36. An adaptive decoding multi-layer video system comprising:a base layer decoder;one or more enhancement layer decoders; anda decoding adaptor connected with the base layer decoder and the one or more enhancement layer decoders, the decoding adaptor controlling operation of the base layer decoder and the one or more enhancement layer decoders.37. The system as recited in claim 36 , further comprising:one or more processing modules providing the one or more enhancement layer decoders with filtered predictions from the base layer decoder, wherein the decoding adaptor further controls operation of the one or more processing modules.38. The system as recited in claim 36 , wherein the decoding adaptor controls operation of the base layer decoder and the one or more enhancement layer decoders based on information gathered from one or more of the base layer decoder and the one or more enhancement layer decoders.39. The system as recited in claim 36 , further comprising a post processing module claim 36 , wherein the decoding adaptor controls operation of the base layer decoder and the one or more enhancement layer decoders based on information gathered from the post processing module.40. The system as recited in claim 39 , wherein the post processing information comprises coding time information related to the base layer decoder and the one or more enhancement layer decoders.41. The system as recited in claim 36 , wherein the decoding adaptor controls operation of the base layer decoder and/or ...

Подробнее
02-05-2013 дата публикации

Systems and Methods for Multi-Layered Image and Video Delivery Using Reference Processing Signals

Номер: US20130106998A1

The present disclosure describes a system and method for multi-layered image and video delivery using reference processing signal. A multi-layered encoder and a multi-layered decoder, both comprising at least one enhancement layer processor that uses reference processing and a base layer processor are provided. A multi-layered encoder and a multi-layered decoder that use reference processing and are capable of frame-compatible 3D video delivery are also described. 126-. (canceled)27. A multi-layered encoding method for image and video delivery , comprising:{'b': 302', '304, 'claim-text': [{'b': 316', '302', '304, 'obtaining at least one first lower resolution image or video frame ( input) by processing the at least one input image or video frame (, ); and'}, {'b': 358', '314, 'obtaining a plurality of base layer encoded bitstream signals () by encoding the at least one first lower resolution image or video frame ( input); and'}], 'base layer processing at least one input image or video frame (, ) through a base layer, said base layer processing comprising{'b': 302', '304, 'claim-text': [{'b': 322', '320', '356, 'claim-text': [{'b': '356', 'the reference processing of a first type comprises interpolating, according to a filtering method, the at least one base layer signal () or the at least one first enhancement layer signal; and'}, {'b': 322', '356, 'the first plurality of reference processing signals () specifies the filtering method of interpolating the at least one base layer signal () or the at least one first enhancement layer signal;'}], 'obtaining a first plurality of reference processing signals () and at least one first reference processed image or video frame () by performing reference processing of a first type on at least one base layer signal () or at least one first enhancement layer signal, according to a first set of rules, wherein, {'b': 326', '320', '302', '304, 'obtaining at least one residual image or video frame () by processing the at least one ...

Подробнее
09-05-2013 дата публикации

Enhancement Methods for Sampled and Multiplexed Image and Video Data

Номер: US20130113884A1

Enhancement methods for sampled and multiplexed image and video data are described. Each component picture is separately processed either after de-multiplexing or on the fly. Processing and de-multiplexing can be combined in a single joint step. The methods apply to both encoding and decoding system and include applications to scalable video coding systems. 117-. (canceled)18. A method to process composite sampled image or video data comprising multiplexed image or video data pertaining to a plurality of image or video components , comprising:de-multiplexing the composite sampled image or video data into a plurality of component pictures;separately processing each component picture; andsampling and multiplexing the separately processed component pictures together.19. A method to process composite sampled image or video data comprising multiplexed image or video data pertaining to a plurality of image or video components or categories , comprising:processing each element of the composite sampled image or video data by taking into account the image or video component or category to which said each element pertains, thus differentiating between processing of composite data pertaining to one image or video component or category and processing of composite data pertaining to another image or video component or category.20. A method to process composite sampled image or video data comprising multiplexed image or video data pertaining to a plurality of image or video components , comprising:de-multiplexing the composite sampled image or video data into a plurality of component pictures while at the same time processing the sampled image or video data, wherein processing is selected from deblocking, denoising, deblurring, deringing, and filtering.21. A method to process composite sampled image or video data comprising multiplexed image or video data pertaining to a plurality of image or video categories , comprising:providing an initial block of existing samples of the same ...

Подробнее
09-05-2013 дата публикации

Systems and Methods for Multi-Layered Frame-Compatible Video Delivery

Номер: US20130114680A1
Принадлежит: Dolby Laboratories Licensing Corp

Multi-layered frame-compatible video delivery is described. Multi-layered encoding and decoding methods, comprising a base layer and at least one enhancement layer with reference processing, are provided. In addition, multi-layered encoding and decoding methods with inter-layer dependencies are described. Encoding and decoding methods that are capable of frame-compatible 3D video delivery are also described.

Подробнее
16-05-2013 дата публикации

Reference Processing Using Advanced Motion Models for Video Coding

Номер: US20130121416A1

Processing a reference picture is described. A reference processing unit enables signaling of parameters such as motion model parameters, interpolation filter parameters, intensity compensation parameters, and denoising filter parameters. Methods for estimating the various parameters are also discussed. Processing improves quality of a reference picture prior to its use for prediction of a subsequent picture and thus improves the prediction. 122-. (canceled)23. A method for processing a reference picture , comprising:i) performing motion estimation on an original picture to be encoded, based on a reconstructed reference picture, and a processed reference picture estimate to provide motion model parameters;ii) further processing the motion estimated picture to obtain a further processed reference picture estimate; anditerating the motion estimation and the further processing to provide a processed reference picture,wherein:the reconstructed reference picture is divided into a plurality of regions,the performing of the motion estimation provides motion model parameters for the plurality of regions, andfor at least one region of the reconstructed reference picture in the plurality of regions, the performing of the motion estimation comprises applying a plurality of motion models to the at least one region of the reconstructed reference picture to provide motion model parameters associated with the at least one region of the reconstructed reference picture, wherein each motion model among the plurality of motion models is applied to the at least one region of the reconstructed reference picture separately, the motion model parameters associated with the at least one region being an average or weighted average of motion model parameters provided by applying each of the plurality of motion models to the at least one region of the reconstructed reference picture separately.24. The method as recited in claim 23 , wherein the further processing comprises further motion ...

Подробнее
23-05-2013 дата публикации

Multiple-Pass Rate Control for Video Coding Applications

Номер: US20130128963A1

Methods and systems for coding applications are described, where multiple coding passes are performed for purposes of rate control. Prediction structures are generated and used by the multiple coding passes to update coding statistics generated prior to the coding passes. The updated coding statistics are used for a subsequent coding operation. 125-. (canceled)26. A method for coding video information in multiple coding passes to control rate of the video information , comprising:providing initial statistics on video information by performing one or more low computational complexity coding passes;performing intermediate coding passes of the video information through one or more asymmetric coding passes to provide updated statistics correcting for inaccuracies introduced by the initial statistics;analyzing the initial statistics and the updated statistics jointly to derive a translation relationship between target statistics and the initial statistics based on the initial statistics and the updated statistics, wherein the target statistics are associated with a further coding pass to be performed; andperforming the further coding pass of the video information at a computational complexity higher than the intermediate coding passes based on the translation relationship between the target statistics and the initial statistics.27. The method as recited in claim 26 , further comprising generating prediction structures based on the video information claim 26 , wherein the asymmetric coding passes code the video information on the basis of the prediction structures.28. The method as recited in claim 26 , wherein the asymmetric coding passes are asymmetric in terms of coding complexity and/or coded frames.29. The method as recited in claim 26 , wherein the initial statistics on video information is provided through the one or more low computational complexity coding passes of the video information at a computational complexity lower than the intermediate coding passes and ...

Подробнее
06-06-2013 дата публикации

Buffered Adaptive Filters

Номер: US20130142262A1

A buffer is established in a video encoder and/or a video decoder. An adaptive filter is buffered in the established buffer. An input video signal is coded using the buffered filter. The filter buffer may be managed, e.g., as to buffer size or capacity, filter placement within the buffer, and buffered default filters. Adaptive filters may be signaled and buffered for multiple reference pictures. 123-. (canceled)24. A method , comprising the steps of:establishing at least one filter buffer in one or more of a video encoder or a video decoder; andbuffering one or more filters in the at least one filter buffer; andcoding an input video signal using the one or more filters in the at least one filter buffer;characterized in that:the one or more buffered filters comprises one or more adaptive filters;the one or more filters in the at least one filter buffer are used to perform one or more of motion interpolation, motion estimation, motion compensation, motion estimation interpolation, motion compensation interpolation, motion prediction interpolation or motion prediction; encoding one or more of a current picture or slice with at least two (2) reference pictures; or', 'decoding one or more of a current picture or slice with at least two (2) reference pictures;, 'one or more ofwherein one or more of the adaptive filters for a first of the at least two reference pictures are signaled independently from at least one of the adaptive filters for at least a second of the at least two reference pictures; andmanaging the at least one filter buffer;wherein the managing step comprises the step of:arranging an order with which the filters are buffered in the at least one filter buffer;wherein the order is arranged by determining the usage count of each filter in the at least one filter buffer, and by sending an ordering syntax in the bitstream from the encoder to the decoder, and by periodically re-ordering the filters in the at least one filter buffer dependent upon the usage count ...

Подробнее
13-06-2013 дата публикации

SPATIOTEMPORAL PREDICTION FOR BIDIRECTIONALLY PREDICTIVE (B) PICTURES AND MOTION VECTOR PREDICTION FOR MULTI-PICTURE REFERENCE MOTION COMPENSATION

Номер: US20130148737A1
Принадлежит: MICROSOFT CORPORATION

Several improvements for use with Bidirectionally Predictive (B) pictures within a video sequence are provided. In certain improvements Direct Mode encoding and/or Motion Vector Prediction are enhanced using spatial prediction techniques. In other improvements Motion Vector prediction includes temporal distance and subblock information, for example, for more accurate prediction. Such improvements and other presented herein significantly improve the performance of any applicable video coding system/logic. 1117.-. (canceled)118. A computing device comprising a processor and memory , wherein the computing device implements a video encoder adapted to perform a method comprising:selectively enabling temporal motion vector (“MV”) prediction for blocks of a current B-frame in a sequence of video frames; if temporal MV prediction is used for the given block of the current B-frame, the predicted MV information for the given block of the current B-frame is derived by performing temporal MV prediction based upon MV information of a block of another frame in the sequence of video frames; and', 'if spatial MV prediction is used for the given block of the current B-frame, (1) the predicted MV information for the given block of the current B-frame is derived by performing spatial MV prediction based upon MV information for one or more surrounding blocks of the current B-frame, and (2) the reference frame for the given block of the current B-frame is selected from among reference frames used for the one or more surrounding blocks of the current B-frame;, 'encoding the blocks of the current B-frame, including, for a given block of the blocks of the current B-frame, encoding the given block of the current B-frame using motion compensation with predicted MV information relative to a reference frame, whereinand wherein at least one of the blocks of the current B-frame is encoded using motion compensation with predicted MV information derived using spatial MV prediction; andoutputting ...

Подробнее
27-06-2013 дата публикации

Method and System for Illumination Compensation and Transition for Video Coding and Processing

Номер: US20130163666A1

Deriving illumination compensation parameters and detection of illumination dominant transitions types for video coding and processing applications is described. Illumination changes such as fade-ins, fade-outs, cross-fades, and flashes are detected. Detection of these illumination changes is then used for weighted prediction to provide for improved illumination compensation. 132-. (canceled)33. A method for fade detection and determination of global or local nature of the fade in a transition from one picture to a next picture of a video signal , the method comprising:providing a plurality of frames of a picture and related prediction reference frames;for each frame and related prediction reference frame; calculating or obtaining one or more intensity related values and one or more color related values in a first color domain, wherein the calculated one or more intensity related values and the one or more color related values comprise color components;for each frame and related prediction reference frame, calculating weighted prediction gains for each calculated color component value in the first color domain;if the weighted prediction gains are all non-negative and substantially similar to each other, then determining that a global transition with zero offset is occurring in a second color domain; and,if the weighted prediction gains are not all non-negative and substantially similar to each other, then determining that at least one of the following is not occurring: a global fading transition; a global fading transition with zero offset; or a global fading transition with zero offset in the second color domain.34. The method of claim 33 , wherein the intensity related values comprise one or more of a luma value or a luminance value.35. The method of claim 33 , wherein the color related values comprise one or more of a chroma value or a chrominance value.36. The method of claim 33 , wherein the first color domain comprises the YCbCr domain and the weighted ...

Подробнее
11-07-2013 дата публикации

Context based Inverse Mapping Method for Layered Codec

Номер: US20130177066A1

Context based inverse mapping methods are provided. An image with pixel values within a lower dynamic range can be mapped to an image with pixel values within a higher dynamic range by utilizing context information associated with pixels in the lower dynamic range image. 1. A method of generating a second pixel based on a first pixel , wherein a first image comprises the first pixel and a second image comprises the second pixel , the first image having a lower dynamic range and the second image having a higher dynamic range , the method comprising:providing a value of the first pixel;generating context information based on at least one pixel neighboring the first pixel; andcomputing a value of the second pixel based on the value of the first pixel and the context information of the at least one pixel to generate the second pixel.2. The method according to claim 1 , wherein the at least one pixel neighboring the first pixel comprises one or more of spatial neighbors and temporal neighbors of the first pixel.3. The method according to claim 1 , wherein the generating of context information comprises providing values of the at least one pixel neighboring the first pixel.4. The method according to claim 3 , further comprising:generating statistical information based on the provided values of the at least one pixel neighboring the first pixel, the statistical information comprises at least one of mean of the provided values, variance of the provided values, texture associated with the provided values, and gradient associated with the provided values, and', 'the computing is based on the statistical information., 'wherein5. The method according to wherein the computing comprises:estimating an inverse mapping function based on the value of the first pixel and the generated context information, the estimated inverse mapping function being adapted to map a value from the lower dynamic range to the higher dynamic range.6. The method according to claim 5 , wherein the ...

Подробнее
18-07-2013 дата публикации

Region Based Asymmetric Coding for 3D Video Compression

Номер: US20130182073A1

Video data that represents two or more views is coded. The regions of the views are asymmetrically processed according to different coding patterns. Techniques that impose asymmetry at the region level, but at the same time may consider overall/average symmetry or significantly reduced asymmetry at the picture/view level are presented. 1. A method for processing video data , comprising:providing video data representing a first view and a second view, the first view and the second view, each view comprising a plurality of regions; andasymmetrically processing the regions of the first view according to a first view processing pattern and asymmetrically processing the regions of the second view according to a second view processing pattern,wherein the first view processing pattern is different from the second view processing pattern.2. The method as recited in claim 1 , wherein the step of asymmetrically processing the regions of the first view and the second view comprises the step of asymmetrically pre-filtering the regions of the first view and the second view.3. The method as recited in claim 1 , wherein the step of asymmetrically processing the regions of the first view and the second view comprises the step of asymmetrically coding the regions of the first view and the second view.4. The method as recited in claim 3 , wherein the step of asymmetrically coding the regions of the first view and the regions of the second view comprises the step of varying a coding parameter from region to region of the first view and varying a coding parameters from region to region of the second view.5. The method as recited in claim 4 , wherein the coding parameter includes one or more of a quantization parameter claim 4 , a coding mode claim 4 , quantization matrices claim 4 , quantization rounding claim 4 , quantization thresholding or application of different Lagrangian parameters.6. The method as recited in claim 5 , wherein the asymmetrical coding of the first view according ...

Подробнее
18-07-2013 дата публикации

Spatiotemporal Metrics for Rate Distortion Optimization

Номер: US20130182971A1

Spatial and temporal metrics are computed for a picture or regions within a picture to determine the impact of coding and quantization on the quality of an encoded picture. Prediction mode decisions and quantization optimization algorithms are used to create a compressed bit stream that minimizes coding artifacts. Pre-processing techniques are also used to suppress coding artifacts that reduce picture quality. 1. A method for reducing artifacts in coded pictures comprising:computing one or more sets of metrics for a picture, wherein each set of metrics are computed based on applying one or more different coding methods;selecting one or more coding methods from the applied one or more different coding methods, wherein selection of the one or more coding methods is based on the computed one or more sets of metrics; andcoding the picture with the selected one or more coding methods.2. The method according to claim 1 , wherein the picture comprises a plurality of picture blocks and computing one or more sets of metrics for the picture comprises computing metrics for one or more of the picture blocks.3. The method according to claim 2 , wherein computing one or more sets of metrics for the picture comprises comparing pixels located at or near a perimeter of a block to pixels located at inner portions of a block.4. The method according to claim 2 , wherein computing one or more sets of metrics for the picture comprises comparing pixels located at or near a boundary of a picture block to pixels located at or near boundaries of one or more adjacent picture blocks.5. The method according to claim 2 , wherein each picture block has a block boundary and computing one or more sets of metrics for the picture comprises computing prediction errors for pixels of a picture block and applying a transform to the computed prediction errors claim 2 , wherein the transform overlaps one or more block boundaries.6. The method according to claim 5 , wherein the prediction errors are ...

Подробнее
25-07-2013 дата публикации

METHODS AND APPARATUSES FOR PROVIDING AN ADAPTIVE REDUCED RESOLUTION UPDATE MODE

Номер: US20130188686A1
Автор: Tourapis Alexandros
Принадлежит: MAGNUM SEMICONDUCTOR, INC.

Methods and apparatuses for applying adaptive reduced resolution update (RRU) processing are disclosed herein. An apparatus may include an encoder configured to receive a video signal and selectively downsample a first component of the video signal in accordance with a first RRU coding mode and a second component of the video signal in accordance with a second RRU coding mode, based on respective types of the first and second components of the video signal. An apparatus may include a decoder configured to receive an encoded bitstream and provide a recovered residual based on the encoded bitstream. The decoder may be configured to selectively upsample a first component of the recovered residual in accordance with a first RRU mode and to selectively upsample a second component of the recovered residual in accordance with a second RRU mode to provide a reconstructed signal based on signaling mechanisms of the encoded bitstream. 1. An apparatus comprising:an encoder configured to receive a video signal and selectively downsample a first component of the video signal in accordance with a first RRU coding mode and a second component of the video signal in accordance with a second RRU coding mode based, at least in part, on the respective types of the first and second components of the video signal.2. The apparatus of claim 1 , wherein the encoder is further configured to selectively downsample the first and second components at a sequence level claim 1 , a frame level claim 1 , a macroblock level claim 1 , or any combination thereof.3. The apparatus of claim 1 , wherein the encoder is further configured to perform motion prediction using full resolution references.4. The apparatus of claim 1 , wherein the encoder is further configured to selectively downsample the first component of the video signal based claim 1 , at least in part claim 1 , on a spatio-temporal analysis of the first component of the video signal.5. An apparatus claim 1 , comprising:a decoder configured ...

Подробнее
01-08-2013 дата публикации

Joint Layer Optimization for a Frame-Compatible Video Delivery

Номер: US20130194386A1

Joint layer optimization for a frame-compatible video delivery is described. More specifically, methods for efficient mode decision, motion estimation, and generic encoding parameter selection in multiple-layer codecs that adopt a reference processing unit (RPU) to exploit inter-layer correlation to improve coding efficiency as described. 1. A method for optimizing coding decisions in a multi-layer layer frame-compatible image or video delivery system comprising one or more independent layers , and one or more dependent layers , the system providing a frame-compatible representation of multiple data constructions , the system further comprising at least one reference processing unit (RPU) between a first layer and at least one of the one or more dependent layers , the first layer being an independent layer or a dependent layer , providing a first layer estimated distortion; and', 'providing one or more dependent layer estimated distortions., 'the method comprising2. The method of claim 1 , wherein the image or video delivery system provides full-resolution representation of the multiple data constructions.3. The method of claim 1 , wherein the RPU is adapted to receive reconstructed region or block information of the first layer.4. The method of claim 1 , wherein the RPU is adapted to receive predicted region or block information of the first layer.5. The method of claim 3 , wherein the reconstructed region or block information input to the RPU is a function of forward and inverse transformation and quantization.6. The method of claim 1 , wherein the RPU uses pre-defined RPU parameters to predict samples for the dependent layer.7. The method of claim 6 , wherein the RPU parameters are fixed.8. The method of claim 6 , wherein the RPU parameters depend on causal past.9. The method of claim 6 , wherein the RPU parameters are a function of the RPU parameters selected from a previous frame in a same layer.10. The method of claim 6 , wherein the RPU parameters are a ...

Подробнее
01-08-2013 дата публикации

Optimized Filter Selection for Reference Picture Processing

Номер: US20130194505A1

Reference processing may be used in a video encoder or decoder to derive reference pictures that are better correlated with a source image to be encoded or decoded, which generally yields better coding efficiency. Methods for filter selection for a reference processing unit adapted for use in a video codec system are discussed. Specifically, methods for filter selection based on performing motion estimation and obtaining distortion/cost information by comparing reference pictures, either processed or non-processed, with the source image to be encoded are discussed. 131-. (canceled)33. The method as recited in claim 32 , wherein the reference picture is encoded after the filtering step.34. The method as recited in claim 32 , wherein the reference picture comprises a base layer reference picture or an inter-layer reference picture claim 32 , wherein an inter-layer comprises a layer from among the one or more enhancement layers.35. The method as recited in claim 32 , wherein the reference picture comprises a spatial reference picture or a temporal reference picture from the enhancement layer.36. The method as recited in claim 32 , wherein the step of providing further comprises processing the enhancement layer source picture claim 32 , wherein the processing removes noise in the enhancement layer source picture.37. The method as recited in claim 32 , wherein the step of providing further comprises processing the enhancement layer source picture claim 32 , wherein the processing involves applying at least one of filtering claim 32 , motion transformation claim 32 , motion compensation claim 32 , illumination compensation claim 32 , scaling claim 32 , inverse and forward tone mapping claim 32 , color format conversion claim 32 , and gamma correction.38. The method as recited in claim 32 , wherein at least one of the filtered reference pictures comprises the reference picture.40. The method as recited in claim 32 , further comprising claim 32 , between the step of ...

Подробнее
15-08-2013 дата публикации

TIMESTAMP-INDEPENDENT MOTION VECTOR PREDICTION FOR PREDICTIVE (P) AND BIDIRECTIONALLY PREDICTIVE (B) PICTURES

Номер: US20130208798A1
Принадлежит: MICROSOFT CORPORATION

Methods and apparatuses are provided for achieving improved video coding efficiency through the use of Motion Vector Predictors (MVPs) for the encoding or decoding of motion parameters within the calculation of the motion information in B pictures and/or P pictures. Certain exemplary methods and apparatuses selectively apply temporal and/or spatial prediction. Rate Distortion Optimization (RDO) techniques are also applied in certain methods and apparatuses to further help improve coding efficiency. 120.-. (canceled)21. A computing device comprising a processor and memory , wherein the computing device implements a video encoder adapted to perform a method comprising:encoding at least part of a current picture of a sequence of pictures using temporal motion vector (“MV”) prediction for direct mode portions; determining a first reference picture for the current direct mode portion as having a first minimum reference picture index among reference picture indices of plural spatially neighboring portions of the current picture for a first reference picture list;', 'determining a first predicted MV for the current direct mode portion using spatial MV prediction, the first predicted MV referencing data associated with the first reference picture, wherein the first predicted MV is based on median values of first MV data for the plural spatially neighboring portions;', 'determining a second reference picture for the current direct mode portion as having a second minimum reference picture index among reference picture indices of the plural spatially neighboring portions for a second reference picture list;', 'determining a second predicted MV for the current direct mode portion using spatial MV prediction, the second predicted MV referencing data associated with the second reference picture, wherein the second predicted MV is based on median values of second MV data for the plural spatially neighboring portions; and', 'performing motion compensation for the current direct ...

Подробнее
29-08-2013 дата публикации

SCALABLE FRAME COMPATIBLE MULTIVIEW ENCODING AND DECODING METHODS

Номер: US20130222539A1

A scalable frame compatible three-dimensional video encoding and decoding system for use in a multiview video coding system is described. A base layer includes low resolution information from a plurality of views while one or more enhancement layers may include high resolution information for at least one of the plurality of views. Interpolation filters are derived based on a combination of low resolution information and high resolution information are discussed. For a given view, sending high resolution information at some times and low resolution information at other times are also described. 121-. (canceled)22. A frame compatible multiview video encoding system adapted to receive information from a plurality of views , comprising:a base layer comprising a base layer encoder, wherein the base layer encoder encodes information from the plurality of views to obtain a first encoded frame compatible image, the first encoded frame compatible image thus comprising a plurality of base layer encoded views;one or more enhancement layers, wherein each enhancement layer is associated with the base layer and each enhancement layer comprises an enhancement layer encoder, wherein at least one view and less than the entirety of views in the plurality of views is encoded by the enhancement layer encoder to obtain a set of encoded view images, each encoded view image being associated with a view among the at least one view and less than the entirety of views; and the filter modes are used to perform interpolation of views in the first encoded frame compatible image and are adapted to be signaled to a decoding system,', 'at least one filter mode is generated based on at least a base layer encoded view among the plurality of base layer encoded views and a corresponding encoded view image among the set of encoded view images, and', 'the at one filter mode is used to perform interpolation of one or more views in the plurality of views., 'a filter generating unit for generating filter ...

Подробнее
29-08-2013 дата публикации

Error Resilient Rate Distortion Optimization for Image and Video Encoding

Номер: US20130223513A1
Принадлежит: Dolby Laboratories Licensing Corp

Error resilient rate distortion optimization (ERRDO) is used for transmitting high quality images and video over constrained bandwidth networks, e.g., in streaming applications. Transmitting high quality images and video by reducing computational complexity is described.

Подробнее
29-08-2013 дата публикации

VIDEO CODING

Номер: US20130223533A1
Принадлежит: MICROSOFT CORPORATION

Improved video coding is described to encode video data within a sequence of video frames. To this end, at least a portion of a reference frame is encoded to include motion information associated with the portion of the reference frame. At least a portion of a predictable frame that includes video data predictively correlated to said portion of said reference frame is defined based on the motion information. At least said portion of the predictable frame is encoded without including corresponding motion information and including mode identifying data. The mode identifying data indicate that the encoded portion of the predictable frame can be directly derived using at least the motion information associated with the portion of the reference frame. 1. A computer-based method for use in encoding video data within a sequence of video frames , the method comprising;encoding at least a portion of at least one reference frame to include motion information associated with said portion of said reference frame;defining at least a portion of at least one predictable frame that includes video data predictively correlated to said portion of said reference frame based on said motion information; andencoding at least said portion of said predictable frame without including corresponding motion information and including mode identifying data that identifies that said portion of said predictable frame can be directly derived using at least said motion information associated with said portion of said reference frame.2. The method as recited in claim 1 , wherein said mode identifying data defines a type of prediction model required to decode said encoded portion of said predictable frame.3. The method as recited in claim 1 , wherein said method generates a plurality of video frames comprising at least one predictable frame selected from a group of predictable frames comprising a P frame and a B frame.4. The method as recited in claim 1 , wherein said portion of said reference frame ...

Подробнее
19-09-2013 дата публикации

Overlapped Rate Control For Video Splicing Applications

Номер: US20130243080A1

Rate control techniques are provided for encoding an input video sequence into a compressed coded bitstream with multiple coding passes. The final coding pass may comprise final splices with non-overlapping frames that do not extend into neighboring final splices. A final splice in the final coding pass may correspond to at least one non-final splice in a non-final coding pass. A non-final splice may have overlapping frames that extend into neighboring final splices in the final coding pass. The overlapping frames in the non-final splice may be used to derive complexity information about the neighboring final splices. The complexity information about the neighboring final splices, as derived from the overlapping frames, may be used to allocate or improve rate control related budgets in encoding the final splice into the compressed coded bitstream in the final coding pass. 1. A method , comprising:dividing an input video sequence into a plurality of final splices to be coded in a final coding pass;performing one or more non-final coding passes before the final coding pass, at least one of the one or more non-final coding passes comprising a non-final splice that corresponds to a final splice in the plurality of splices to be coded in the final coding pass, and the non-final splice comprising more frames than frames in the final splice;allocating one or more rate control related budgets for the final splice based on information derived from the non-final splice in the one or more non-final coding passes; andcoding the final splice in the final coding pass using the one or more rate control related budgets.2. The method as recited in claim 1 , wherein the one or more rate control related budgets include one or more of a quantization parameter claim 1 , a total number of header bits allocated to a splice-level bit stream coded from the final splice claim 1 , or a total number of texture bits allocated to the splice-level bit stream coded from the final splice.3. The ...

Подробнее
02-01-2014 дата публикации

Bitdepth and Color Scalable Video Coding

Номер: US20140003527A1
Автор: Tourapis Alexandros

Methods for scalable video coding are described. Such methods can be used to deliver video contents in Low Dynamic Range (LDR) and/or one color format and then converting the video contents to High Dynamic Range (HDR) and/or a different color format, respectively, in block or macroblock levels. 136-. (canceled)37. A method of mapping input video data from a first layer to a second layer , the method comprising:selecting one or more inter-layer prediction methods from among a plurality of inter-layer prediction methods for each video block or macroblock of a plurality of video blocks or macroblocks on a first layer, each of the video blocks or macroblocks comprising a portion of input video data, wherein, at least for one of the plurality of video blocks or macroblocks, more than one inter-layer prediction methods from among the plurality of inter-layer prediction methods are selected;mapping each video block or macroblock of the first layer to a second layer by applying, for each video block or macroblock, the selected one or more inter-layer prediction methods; andselecting by a selector the selected prediction methods; andsignaling the selected prediction methods within a parameter set or a coding unit.38. The method of claim 37 , wherein a video block or macroblock of the first layer has a low dynamic range; wherein a video block or macroblock of the second layer has a high dynamic range; wherein the plurality of prediction methods is a plurality of inverse tone mapping methods.39. The method of claim 38 , wherein an inverse tone mapping method is further selected from one or more of:a) linear scaling and clipping;b) linear interpolation;c) lookup table mapping;d) color formation;e) N-th order polynomials; andf) splines.40. The method of claim 37 , wherein the parameter set or coding unit is a Sequence Parameter Set (SPS) claim 37 , a Picture Parameter Set (PPS) claim 37 , or a Slice Header.41. The method of claim 37 , wherein the parameter set or coding unit is ...

Подробнее
02-01-2014 дата публикации

Pre-Processing for Bitdepth and Color Format Scalable Video Coding

Номер: US20140003528A1
Автор: Tourapis Alexandros

Methods for scalable video coding are described. Such methods can be used to deliver video contents in Low Dynamic Range (LDR) and/or one color format and then converting the video contents to High Dynamic Range (HDR) and/or a different color format, respectively, while pre-processing video content. 130-. (canceled)31. A method of processing input video data comprising:receiving a first input of video data;accessing a second input of video data, wherein the second input of video data is generated based on the first input and has a lower dynamic range than the first input;pre-processing the first input video data with a first Motion Compensated Temporal Filter (MCTF) to generate a first pre-processed signal;pre-processing the second input video data with a second Motion Compensated Temporal Filter (MCTF) to generate a second pre-processed signal;encoding the second pre-processing signal using a base layer encoder to generate a base layer stream; andencoding the first pre-processing signal using an enhancement layer encoder to generate an enhancement layer stream;32. The method of claim 31 , wherein the second input is generated by tone mapping and/or color formatting the first input video data.33. The method of claim 31 , wherein encoding the first pre-processed signal using an enhancement layer encoder is based on both the first pre-processed signal and the second pre-processed signal.34. The method of wherein the base layer stream and the enhancement layer stream are multiplexed to generate a combined output bitstream.35. A method of processing input video data claim 31 , the method comprising:providing a base layer and at least one enhancement layer;providing a pre-processing unit, the pre-processing unit including at least a Motion Compensated Temporal Filter (MCTF) and a tone mapping and/or color format conversion function block, wherein the tone mapping and/or color format conversion function block is arranged downstream the Motion Compensated Temporal Filter; ...

Подробнее
13-02-2014 дата публикации

Entropy coding techniques and protocol to support parallel processing with low latency

Номер: US20140044194A1
Принадлежит: Apple Inc

In a communication system, parallel encoding and decoding of serially-coded data occurs in a manner that supports low latency communication. A plurality of data items may be coded as serially-coded data sequences and a transmission sequence may be built from them. An index table may be built having a plurality of entries representing respective start points of the serially-coded data sequences within the transmission sequence. The transmission sequence may be transmitted to a channel and, thereafter, the index table may be transmitted. Latencies otherwise involved in inserting an index table into the beginning of a transmission sequence may be avoided.

Подробнее
13-03-2014 дата публикации

VIDEO DEBLOCKING FILTER STRENGTH DERIVATION

Номер: US20140072043A1
Принадлежит: Apple Inc.

Codecs may be modified to consider weighting and/or illumination compensation parameters when determining a deblocking filter strength that is to be applied. These parameters may be useful for recording illumination changes, such as fades, cross-fades, flashes, or light source changes, which allows these illumination changes to displayed during playback using the same reference frame data which different weighting and/or illumination compensation parameters applied. In different instances, the parameters may be considered when setting a deblocking filter strength to ensure that these effects are properly displaying during playback while minimizing the appearance of blocking artifacts. 1. A method for configuring a deblocking filter to reduce banding artifacts comprising:comparing a weighted prediction parameter of a video codec inter-prediction process from a reference index in a plurality of blocks using a processing device;when the compared weighted prediction parameter in the blocks is different, setting a deblocking filter strength of the blocks to a first value; calculating a difference between motion vectors of the respective blocks in a horizontal direction and a vertical direction;', 'when the difference in at least one of the directions is greater than or equal to a threshold, setting the deblocking filter strength of the blocks to a second value; and', 'when the difference in both directions is less than the threshold, setting the deblocking filter strength of the blocks to a third value., 'when the weighted prediction parameter in the blocks is similar2. The method of claim 1 , wherein the first value and the second value are equal and greater than the third value claim 1 , the first value and the second value indicating filtering should be applied to the block claim 1 , and the third value indicating that filtering should be skipped for the blocks.3. The method of claim 2 , further comprising claim 2 , when at least one of the blocks has at least one non ...

Подробнее
13-03-2014 дата публикации

Video display preference filtering

Номер: US20140072057A1
Принадлежит: Apple Inc

Video coding schemes may include one or more filters to reduce coding artifacts and improve video quality. These filters may be applied to decode video data in a predetermined sequence. The output from one or more of these filters may be selected for different images, blocks, or sets of video data and then copied and/or routed to a display or a buffer storing reference data that is used to decode other video data in a data stream. Providing the ability to select which filter output is used for display and as a reference may result in better video quality for multiple types of video data. The filters that are selected for display and for reference may be different and may vary for different images, blocks, and data sets.

Подробнее
03-04-2014 дата публикации

ENTROPY CODING TECHNIQUES AND PROTOCOL TO SUPPORT PARALLEL PROCESSING WITH LOW LATENCY

Номер: US20140092987A1
Принадлежит: Apple Inc.

In a communication system, parallel encoding and decoding of serially-coded data occurs in a manner that supports low latency communication. A plurality of data items may be coded as serially-coded data sequences and a transmission sequence may be built from them. An index table may be built having a plurality of entries representing respective start points of the serially-coded data sequences within the transmission sequence. The transmission sequence may be transmitted to a channel and, thereafter, the index table may be transmitted. Latencies otherwise involved in inserting an index table into the beginning of a transmission sequence may be avoided. 1. A method , comprising:coding a plurality of data items as serially-coded data sequences,building a transmission sequence from the serially-coded data sequences,building a table having a plurality of entries representing decoding parameters to be applied during decoding of the respective coded data sequences, andtransmitting the transmission sequence and the table in a channel, wherein the transmission sequence precedes the index table in transmission order.2. The method of claim 1 , wherein the decoding parameters include deblocking filter information for the respective coded data sequences.3. The method of claim 1 , wherein the decoding parameters include checksum data for the respective coded data sequences.4. The method of claim 1 , wherein the decoding parameters include signatures for the respective coded data sequences.5. The method of claim 1 , wherein the transmission sequence and the table are transmitted in a common Network Adaptation Layer Unit.6. The method of claim 1 , wherein the table is transmitted in a Network Adaptation Layer (NAL) Unit separate from NAL Unit(s) in which the transmission sequence is transmitted.7. The method of claim 1 , wherein the table is transmitted in a Supplemental Enhancement Information message.8. A method claim 1 , comprising:receiving a transmission sequence and a table ...

Подробнее
06-01-2022 дата публикации

Point Cloud Compression Using Non-Cubic Projections and Masks

Номер: US20220005228A1
Принадлежит: Apple Inc

A system comprises an encoder configured to compress attribute information and/or spatial for a point cloud and/or a decoder configured to decompress compressed attribute and/or spatial information for the point cloud. To compress the attribute and/or spatial information, the encoder is configured to convert a point cloud into an image based representation. Also, the decoder is configured to generate a decompressed point cloud based on an image based representation of a point cloud.

Подробнее
07-01-2021 дата публикации

Point Cloud Compression with Supplemental Information Messages

Номер: US20210006833A1
Принадлежит: Apple Inc.

A system comprises an encoder configured to compress attribute information and/or spatial for a point cloud and/or a decoder configured to decompress compressed attribute and/or spatial information for the point cloud. To compress the attribute and/or spatial information, the encoder is configured to convert a point cloud into an image based representation. Also, the decoder is configured to generate a decompressed point cloud based on an image based representation of a point cloud. Additionally, an encoder is configured to signal and/or a decoder is configured to receive a supplementary message comprising volumetric tiling information that maps portions of 2D image representations to objects in the point. In some embodiments, characteristics of the object may additionally be signaled using the supplementary message or additional supplementary messages. 1. A non-transitory computer-readable medium storing program instructions , that when executed on or across one or more processors , cause the one or more processors to: generate, for the respective ones of the respective sets of points, respective attribute patch images and respective depth patch images comprising attribute information or geometry information for the respective sets of points projected onto respective patch planes;', 'pack the generated attribute patch images and the generated depth patch images in one or more 2D image frames;', 'video encode the one or more 2D image frames; and', 'encode one or more supplementary messages comprising one or more entries, wherein the one or more entries comprise information indicating portions of the one or more 2D image frames that comprise attribute and depth patch images that correspond to an object in the point cloud being encoded., 'compress a point cloud, wherein to compress the point cloud, the program instructions cause the one or more processors to, for respective ones of a plurality of respective sets of points of the point cloud2. The non-transitory ...

Подробнее
02-01-2020 дата публикации

COMPRESSION WITH MULTI-LEVEL ENCODING

Номер: US20200007867A1
Принадлежит: Apple Inc.

A system comprises an encoder configured to compress images, such as image frames comprising attribute information and/or spatial for a point cloud and/or an occupancy map for the point cloud. Also, a system includes a decoder configured to decompress compressed image frames, such as image frames comprising compressed attribute and/or spatial information for the point cloud or an occupancy map for the point cloud. Additionally, the encoder may map N-bit data to M-bit code words, where M is less than N. Alternatively the encoder may map N-bit data to M-bit code words, where M is greater than N. In a similar manner, a decoder may map the M-bit code words back to the N-bit data. 1. A system , comprising:an encoder configured to encode an image with pixel values comprising N-bit pixel values into one or more image frames with M-bit pixel values, determine whether a bit-depth of the N-bit pixel values of pixels of the image being encoded exceeds an image frame bit-depth supported by the encoder; and', [ round the N-bit pixel value to a pixel value expressed using an M-bit pixel value supported by the encoder, where M is less than N; and', 'determine an error value between the N-bit pixel value of the image being encoded and a rounded pixel value expressed using the M-bit pixel value;, 'for each N-bit pixel value of the image being encoded, 'encode a first image frame for the image being encoded, wherein the rounded pixel values expressed using the M-bit pixel values are encoded for pixels in the first image frame at pixel locations corresponding to the pixels of the image being encoded; and', 'encode a second image frame for the image being encoded, wherein the determined error values are encoded for pixels in the second image frame at pixel locations corresponding to the pixels of the image being encoded., 'in response to determining the bit-depth of the N-bit pixel values of the pixels of the image being encoded exceeds the image frame bit-depth supported by the ...

Подробнее
14-01-2021 дата публикации

POINT CLOUD COMPRESSION WITH MULTI-LAYER PROJECTION

Номер: US20210012536A1
Принадлежит: Apple Inc.

A system comprises an encoder configured to compress attribute information and/or spatial for a point cloud and/or a decoder configured to decompress compressed attribute and/or spatial information for the point cloud. To compress the attribute and/or spatial information, the encoder is configured to convert a point cloud into an image based representation. Also, the decoder is configured to generate a decompressed point cloud based on an image based representation of a point cloud. The encoder is configured project the point cloud on to patch planes to compress the point cloud, and supports multiple layered patch planes. For example, some point clouds may have a depth, and points at different depths may be assigned to different layered patch planes. 120.-. (canceled)21. A non-transitory computer-readable medium storing program instructions that , when executed by one or more processors , cause the one or more processors to: a patch image comprising a first set of points of the point cloud corresponding to a patch projected onto a patch plane; and', 'another patch image comprising another set of points of the point cloud, at a different layer of the point cloud, corresponding to the patch projected on the patch plane;, 'receive encoded point cloud information comprisingdecode the encoded point cloud information; andgenerate a reconstructed point cloud based on the decoded point cloud information.22. The non-transitory computer-readable medium of claim 21 , wherein the points of the other set of points comprise points that are located at respective same pixel locations in the patch plane as points of the first set of points.23. The non-transitory computer-readable medium of claim 21 , wherein the patch image comprising the other set of points of the point cloud at the different layer has been down-scaled relative to the patch image comprising the first set of points of the point cloud.24. The non-transitory computer-readable medium of claim 23 , wherein the program ...

Подробнее
12-01-2017 дата публикации

SYSTEMS AND METHODS FOR MULTI-LAYERED IMAGE AND VIDEO DELIVERY USING REFERENCE PROCESSING SIGNALS

Номер: US20170013277A1
Принадлежит:

The present disclosure describes a system and method for multi-layered image and video delivery using reference processing signal. A multi-layered encoder and a multi-layered decoder, both comprising at least one enhancement layer processor that uses reference processing and a base layer processor are provided. A multi-layered encoder and a multi-layered decoder that use reference processing and are capable of frame-compatible 3D video delivery are also described. 1. A decoding method for multi-layered frame-compatible video delivery with one or more processors , the method comprising:receiving a coded bitstream comprising a coded base layer, a coded first enhancement layer, and a coded second enhancement layer;base layer processing of the coded base layer using a base layer processor to generate at least one frame-compatible base layer decoded image, wherein the at least one frame-compatible base layer decoded image comprises a base layer image of a first view in a first spatial resolution and a base layer image of a second view in the first spatial resolution multiplexed together in a multiplexing pattern;a first enhancement layer processing of the coded first layer bitstream to generate a first decoded residual image for the first view at a second spatial resolution higher than the first spatial resolution;a second enhancement layer processing of the coded second layer bitstream to generate a second decoded residual image for the second view at the second spatial resolution;a first reference processing to generate a first reference image at the second spatial resolution based on a decoded base layer image of the first view and a first filtering method specified in the coded bitstream;a second reference processing to generate a second reference image at the second spatial resolution based on a decoded base layer image of the second view and a second filtering method specified in the coded bitstream;performing disparity compensation using the first reference image ...

Подробнее
15-01-2015 дата публикации

Spatiotemporal prediction for bidirectionally predictive (b) pictures and motion vector prediction for multi-picture reference motion compensation

Номер: US20150016527A1
Принадлежит: Microsoft Corp

Several improvements for use with Bidirectionally Predictive (B) pictures within a video sequence are provided. In certain improvements Direct Mode encoding and/or Motion Vector Prediction are enhanced using spatial prediction techniques. In other improvements Motion Vector prediction includes temporal distance and subblock information, for example, for more accurate prediction. Such improvements and other presented herein significantly improve the performance of any applicable video coding system/logic.

Подробнее
09-01-2020 дата публикации

DEVICE AND METHOD FOR FAST BLOCK-MATCHING MOTION ESTIMATION IN VIDEO ENCODERS

Номер: US20200014952A1
Принадлежит:

A solution is provided to estimate motion vectors of a video. A multistage motion vector prediction engine is configured to estimate multiple best block-matching motion vectors for each block in each video frame of the video. For each stage of the motion vector estimation for a block of a video frame, the prediction engine selects a test vector form a predictor set of test vectors, computes a rate-distortion optimization (RDO) based metric for the selected test vector, and selects a subset of test vectors as individual best matched motion vectors based on the RDO based metric. The selected individual best matched motion vectors are compared and a total best matched motion vector is selected based on the comparison. The prediction engine selects iteratively applies one or more global matching criteria to the selected best matched motion vector to select a best matched motion vector for the block of pixels. 1receive a video comprising a plurality of video frames, each frame comprising a plurality of blocks; selecting a center position motion vector as a best matched motion vector by iteratively checking a set of test motion vectors based on a first rate-distortion optimization (RDO)-based metric, the set of test motion vectors expected to include highly reliable predictors based on at least one of a priori knowledge of a video including the video frame and a priori knowledge of a plurality of video sequences stored in a database, the set of test motion vectors comprising one or more of a zero-motion vector, a motion vector predictor (MVP), and one or more individual motion vectors from neighboring blocks adjacent to the block, the MVP being a median of a plurality of motion vectors for a plurality of neighboring blocks adjacent to the block; and', [ determining whether a candidate total best matched motion vector for the iteration meets one or more first adaptive threshold criteria based on a second RDO-based metric; and', 'responsive to the candidate total best ...

Подробнее
09-01-2020 дата публикации

POINT CLOUD COMPRESSION WITH MULTI-RESOLUTION VIDEO ENCODING

Номер: US20200014953A1
Принадлежит: Apple Inc.

A system comprises an encoder configured to compress attribute and/or spatial information for a point cloud and/or a decoder configured to decompress compressed attribute and/or spatial information for the point cloud. To compress the attribute and/or spatial information, the encoder is configured to convert a point cloud into an image based representation. Also, the decoder is configured to generate a decompressed point cloud based on an image based representation of a point cloud. In some embodiments, an encoder performs downscaling of an image frame prior to video encoding and a decoder performs upscaling of an image frame subsequent to video decoding. 1. A system , comprising: determine, for the point cloud, a plurality of patches each corresponding to 3D portions of the point cloud projected onto respective 2D planes;', 'generate an attribute patch image and a geometry patch image for respective ones of the patches;', 'pack the generated attribute patch images and the generated geometry patch images into one or more 2D image frames;', 'down-scale at least one of the one or more 2D image frames, wherein downscaling the at least one 2D image frame reduces a quantity of pixels of the attribute patch images or geometry patch images packed into the at least one 2D image frame; and', 'video encode the one or more 2D image frames including the at least one down-scaled 2D image frame., 'an encoder configured to compress data representing a point cloud, wherein to compress the data representing the point cloud, the encoder is configured to2. The system of claim 1 , wherein the one or more 2D image frames comprises:a 2D attribute image frame comprising respective ones of the attribute patch images; anda 2D geometry image frame comprising respective ones of the geometry patch images,wherein the down-scaling of the at least one of the one or more 2D image frames causes the 2D attribute image frame or the 2D geometry image frame to be down-scaled by a different respective ...

Подробнее
18-01-2018 дата публикации

VIDEO COMPRESSION AND TRANSMISSION TECHNIQUES

Номер: US20180020220A1
Принадлежит:

Embodiments feature families of rate allocation and rate control methods that utilize advanced processing of past and future frame/field picture statistics and are designed to operate with one or more coding passes. At least two method families include: a family of methods for a rate allocation with picture look-ahead; and a family of methods for average bit rate (ABR) control methods. At least two other methods for each method family are described. For the first family of methods, some methods may involve intra rate control. For the second family of methods, some methods may involve high complexity ABR control and/or low complexity ABR control. These and other embodiments can involve any of the following: spatial coding parameter adaptation, coding prediction, complexity processing, complexity estimation, complexity filtering, bit rate considerations, quality considerations, coding parameter allocation, and/or hierarchical prediction structures, among others. 1. (canceled)2. A method for decoding a bitstream , the method comprising: wherein image frames of the second temporal scalability level are not used as a reference for motion-compensated prediction of image frames of the first temporal scalability level,', 'wherein at least one frame of the second temporal scalability level is signaled in the bitstream as a disposable frame, and at least one frame of the second temporal scalability level is not signaled in the bitstream as a disposable frame, and', 'wherein at least one frame of the second temporal scalability level is an I-coded picture;, 'receiving, at a decoder comprising one or more processing devices, image frames associated with at least a first temporal scalability level and a second temporal scalability level,'}discarding, without decoding, all frames of the second temporal scalability level; anddecoding frames of the first temporal scalability level.3. The method of wherein the at least one frame of the second temporal scalability level that is not ...

Подробнее
16-01-2020 дата публикации

BIT STREAM STRUCTURE FOR COMPRESSED POINT CLOUD DATA

Номер: US20200021847A1
Принадлежит: Apple Inc.

A system comprises an encoder configured to compress attribute information and/or spatial information for a point cloud and/or a decoder configured to decompress compressed attribute and/or spatial information for the point cloud. To compress the attribute and/or spatial information, the encoder is configured to convert a point cloud into an image based representation. Also, the decoder is configured to generate a decompressed point cloud based on an image based representation of a point cloud. In some embodiments, a bit stream structure may be used to communicate compressed point cloud data. The bit stream structure may include point cloud compression network abstraction layer (PCCNAL) units that enable use of groups of frames (GOFs), frame, and sub-frame signaling of patch information. Such a bit stream structure may permit low delay streaming and random access reconstruction of point clouds amongst other applications. 1. A system comprising:one or more sensors configured to capture a plurality of points that make up a point cloud, wherein respective ones of the points comprise spatial information for the point and attribute information for the point; and determine, for the point cloud, a plurality of patches each corresponding to portions of the point cloud, and', 'pack generated patch images for the determined patches into one or more image frames; and, 'an encoder configured to compress the point cloud, wherein to compress the point cloud, the encoder is configured toencode the one or more image frames into a bit stream comprising point cloud compression network abstraction layer (PCCNAL) units, wherein the PCCNAL units indicate locations of one or more of the patch images in the bit stream.2. The system of claim 1 , wherein at least one of the PCCNAL units indicates a relationship between a patch image in a first image frame of the bit stream and another patch image in another image frame of the bit stream.3. The system of claim 2 , wherein the relationship ...

Подробнее
16-01-2020 дата публикации

HIERARCHICAL POINT CLOUD COMPRESSION

Номер: US20200021856A1
Принадлежит: Apple Inc.

A system comprises an encoder configured to compress attribute information for a point cloud and/or a decoder configured to decompress compressed attribute information for the point cloud. To compress the attribute information, multiple levels of detail are generated based on an ordering of the points according to a space filling curve and attribute values are predicted. The attribute values may be predicted simultaneously while points are being assigned to different levels of detail. A decoder follows a similar prediction process based on level of details. Also, attribute correction values may be determined to correct predicted attribute values and may be used by a decoder to decompress a point cloud compressed using level of detail attribute compression. In some embodiments, attribute correction values may take into account an influence factor of respective points in a given level of detail on attributes in other levels of detail. 1. A system , comprising: organize a plurality of points of a point cloud into an order according to a space filling curve based on respective spatial positions of the plurality of points of the point cloud in 3D space;', 'assign an attribute value to at least one point of a first level of detail for the point cloud, wherein points of the point cloud to be included in the first level of detail are selected based, at least in part, on their respective positions in the order according to the space filling curve;', 'for respective points of the other points of the first level of detail, and points of one or more additional levels of detail for the point cloud, determine a predicted attribute value for the respective point based on predicted or assigned attributes values for neighboring points in a same level of detail as the point for which an attribute value is being predicted, wherein points to be included in the one or more additional levels of detail are selected based, at least in part, on their respective positions in the order ...

Подробнее
24-04-2014 дата публикации

Systems and Methods for Multi-Layered Frame-Compatible Video Delivery

Номер: US20140111614A1
Принадлежит: Dolby Laboratories Licensing Corp

Multi-layered frame-compatible video delivery is described. Multi-layered encoding and decoding methods, comprising a base layer and at least one enhancement layer with reference processing, are provided. In addition, multi-layered encoding and decoding methods with inter-layer dependencies are described. Encoding and decoding methods that are capable of frame-compatible 3D video delivery are also described.

Подробнее
25-01-2018 дата публикации

Speedup Techniques for Rate Distortion Optimized Quantization

Номер: US20180027245A1

Techniques for selecting a coding mode for an image coding process are described. Coding modes can be selected through a coding mode transition state machine, a re-quantization process, selection of an optimal transform size, by skipping some quantization parameters, or by performing motion search. 1. A method of selecting a coding mode and a quantization parameter for an image coding unit to be coded and quantized in an image coding process , comprising:(a) performing a first quantization process having a first type of quantization technique during evaluation of M possible coding modes;{'sub': 1', '1, '(b) based on the first quantization process, selecting Mbest coding modes according to a predefined coding criterion, wherein M>M≧1;'}{'sub': '1', '(c) after steps (a) and (b) have been performed for each of the M possible coding modes, performing a second quantization process that has a second type of quantization technique for the Mbest coding modes; and'}(d) selecting a best coding mode based on the results of the first quantization process and the second quantization process;wherein the first quantization process has the first type of quantization technique, wherein the first type of quantization technique comprises a faster type of quantization than the second quantization type;wherein the second quantization process has the second type of quantization technique, which comprises a slower quantization type than the first quantization type; andwherein the first quantization process and the second quantization process are repeated for each quantization parameter.2. The method as recited in claim 1 , wherein information about the positions and magnitudes of quantized transform coefficients of one or more quantization parameters is used during the quantization processes of subsequent quantization parameters.3. The method as recited in claim 1 , wherein the second quantization process is applied to a single best coding mode among the M possible coding modes claim 1 , ...

Подробнее
28-01-2021 дата публикации

Point Cloud Geometry Compression Using Octrees and Binary Arithmetic Encoding with Adaptive Look-Up Tables

Номер: US20210029383A1
Принадлежит: Apple Inc.

An encoder is configured to compress point cloud geometry information using an octree geometric compression technique that utilizes a binary arithmetic encoder, a look-ahead table, a cache, and a context selection process, wherein encoding contexts are selected based, at least in part, on neighborhood configurations. In a similar manner, a decoder is configured to decode compressed point cloud geometry information utilizing a binary arithmetic encoder, a look-ahead table, a cache, and a context selection process. 120-. (canceled)21. One or more non-transitory , computer-readable storage media , storing program instructions that , when executed on or across one or more computing devices , cause the one or more computing devices to:decode occupancy symbols for divisions of an encoded point cloud, wherein the encoded point cloud has been encoded via an octree geometrical compression technique, determine whether a first bit is set indicating that the given occupancy symbol is included in a look-up table,', 'wherein if the first bit indicates the given occupancy symbol is included in the look-up table, the given occupancy symbol is read from the look-up table based on an index value included in the encoded point cloud, wherein the index value corresponds to the given occupancy symbol in the look-up table;', 'determine, if the first bit is not set, whether another bit is set indicating that the given occupancy symbol is included in a cache,', 'wherein if the other bit indicates the given occupancy symbol is included in the cache, the given occupancy symbol is read from the cache based on an index value included in the encoded point cloud, wherein the index value corresponds to the given occupancy symbol in the cache; and', 'otherwise decode a binary representation of the given occupancy symbol included in the encoded point cloud., 'one or more computing devices to, 'wherein to decode a given one of the occupancy symbols, the program instructions, when executed on or ...

Подробнее
01-05-2014 дата публикации

Encoder Optimization of Stereoscopic Video Delivery Systems

Номер: US20140118491A1

Controlling a feature of an encoding process for regions of an image pattern representing more than one image when the regions include an amount of disparity in the represented images that would result in cross-contamination between the represented images if encoded with the feature. The control may be, for example, any of, turning the encoding feature off, using the encoding feature less often than when encoding an image pattern representing a single image, negatively biasing the encoding feature, and enabling the encoding feature for regions determined to have zero or near zero disparity and disabling the feature for all other regions. The represented images comprise, for example, any of a stereoscopic view, multiple stereoscopic views, multiple views of a same scene, and multiple unrelated views. 115-. (canceled)16. A method of encoding images comprising:receiving an image with an image pattern comprising a checkerboard arrangement of pixels from two channels of a stereoscopic view;classifying the disparity of regions in the image as being one of high disparity, low disparity and zero disparity, wherein the disparity is determined by at least one of bandpass filtering the image patterns and computing distortion, wherein the distortion is computed by comparing original pre-image pattern images to the images represented in the image pattern after being decoded, wherein the amount of disparity comprises at least one of an amount of estimated motion in corresponding regions of the images, and an amount of illumination change in corresponding regions of the images;turning off or negatively biasing an encoding feature of an encoding process for regions of an image pattern based on the disparity classification, when the regions include an amount of disparity in the represented images of the same frame such that the amount of disparity in the represented images would result in cross-contamination between the represented images if encoded with the encoding feature.17. The ...

Подробнее
01-05-2014 дата публикации

RATE-DISTORTION OPTIMIZERS AND OPTIMIZATION TECHNIQUES INCLUDING JOINT OPTIMIZATION OF MULTIPLE COLOR COMPONENTS

Номер: US20140119454A1
Принадлежит: MAGNUM SEMICONDUCTOR, INC.

Examples of encoders and video encoding are described that include optimizers and techniques for optimizing syntax elements such as transform coefficients. In some examples, multiple color components of a video signal may be jointly optimized by employing a cost calculation using a combination of distortion and/or rate metrics for multiple color components. In some examples, a color transformation may occur and the optimization may take place in a different color domain than encoding. In some examples, distortion metrics used in the cost calculations performed by optimizers are based on structural similarity index. 1. A video encoder comprising:a transform configured to transform a residual representation of a video signal to transform coefficients;an optimizer configured to receive the transform coefficients and provide optimized coefficients using a cost calculation based on rate and distortion, and wherein the cost calculation includes a combination of distortion metrics due to multiple color components of the video signal; andan entropy encoder configured to encode the video signal using the optimized transform coefficients.2. The video encoder of claim 1 , wherein the entropy encoder is configured to encode the video signal in a first color domain for display in a second color domain claim 1 , and wherein the combination of distortion metrics further includes distortion metrics calculated in the second color domain.3. The video encoder of claim 2 , wherein the first color domain is YUV and the second color domain is RGB.4. The video encoder of claim 2 , wherein the optimizer comprises a color transform configured to transform the transform coefficients from the first color domain to the second color domain.5. The video encoder of claim 4 , wherein the optimizer further comprises an interpolator configured to interpolate at least portions of the video signal and provide the interpolated portions to the color transform.6. The video encoder of claim 1 , wherein at ...

Подробнее
04-02-2021 дата публикации

Adaptive Transfer Function for Video Encoding and Decoding

Номер: US20210037246A1
Принадлежит:

A video encoding and decoding system that implements an adaptive transfer function method internally within the codec for signal representation. A focus dynamic range representing an effective dynamic range of the human visual system may be dynamically determined for each scene, sequence, frame, or region of input video. The video data may be cropped and quantized into the bit depth of the codec according to a transfer function for encoding within the codec. The transfer function may be the same as the transfer function of the input video data or may be a transfer function internal to the codec. The encoded video data may be decoded and expanded into the dynamic range of display(s). The adaptive transfer function method enables the codec to use fewer bits for the internal representation of the signal while still representing the entire dynamic range of the signal in output. 120-. (canceled)22. The system as recited in claim 21 , wherein the range of luminance values extracted from the N-bit video data represents an effective dynamic range of the human visual system for the N-bit video data.23. The system as recited in claim 21 , wherein the video decoder is further configured to obtain format metadata for the encoded C-bit video data claim 21 , wherein the format metadata includes an indication of the range of luminance values extracted from the N-bit video data.24. The system as recited in claim 23 , wherein the format metadata is embedded in the encoded C-bit video data claim 23 , and wherein claim 23 , to obtain the format metadata claim 23 , the video decoder extracts the format metadata from the decoded C-bit video data.25. The system as recited in claim 23 , wherein the format metadata further includes an indication of a transfer function used to represent the C-bit video data.26. The system as recited in claim 23 , wherein the video decoder is configured to perform said obtain and said expand for each of one or more video frames or for each of one or more ...

Подробнее
07-02-2019 дата публикации

SPATIOTEMPORAL PREDICTION FOR BIDIRECTIONALLY PREDICTIVE (B) PICTURES AND MOTION VECTOR PREDICTION FOR MULTI-PICTURE REFERENCE MOTION COMPENSATION

Номер: US20190045216A1
Принадлежит: Microsoft Technology Licensing, LLC

Several improvements for use with Bidirectionally Predictive (B) pictures within a video sequence are provided. In certain improvements Direct Mode encoding and/or Motion Vector Prediction are enhanced using spatial prediction techniques. In other improvements Motion Vector prediction includes temporal distance and subblock information, for example, for more accurate prediction. Such improvements and other presented herein significantly improve the performance of any applicable video coding system/logic. 1117.-. (canceled)118. A computing device comprising one or more processing units , volatile memory , non-volatile memory , and storage , the non-volatile memory and/or storage having stored therein computer-executable instructions for causing the computing device , when programmed thereby , to perform operations comprising:receiving encoded data for a current frame in a sequence of video frames; and [ determining a first MV predictor from a first surrounding block of the current frame, wherein the first MV predictor uses a second reference picture that has a second temporal distance from the current frame;', 'scaling the first MV predictor according to the first temporal distance and the second temporal distance;', 'determining a second MV predictor from a second surrounding block of the current frame, wherein the second MV predictor uses a third reference frame that has a third temporal distance from the current frame; and', 'scaling the second MV predictor according to the first temporal distance and the third temporal distance; and, 'determining a predicted motion vector (“MV”) for the given block using multiple MV predictors from the current frame, including, 'performing motion compensation for the given block relative to the first reference frame to determine a motion-compensated prediction for the given block., 'decoding the current frame, including, for a given block of the current frame, wherein the given block uses a first reference frame that has a first ...

Подробнее
18-02-2021 дата публикации

Adaptive Bilateral Filtering Using Look-Up Tables

Номер: US20210051320A1
Принадлежит:

A system comprises an encoder configured to compress video data. The encoder includes an adaptive bilateral filter that uses look-up tables. The encoder may encode one or more adaptive adjustment factors to be used by a decoder to select or adjust look-up tables used to decode the compressed video data. 2. The system of claim 1 , wherein the encoder is configured to sum the approximated offset values and apply a shift operator based on a number of neighboring pixels being used to the determine the bilateral filter offset value claim 1 , andwherein the encoder is configured to take an absolute value of the summed approximated offset values prior to applying the shift operator and then apply a sign, positive or negative, associated with the summed approximated offset values after applying the shift operator.3. The system of or claim 1 , wherein the one or more characteristics of the respective portion used to determine the adaptive adjustment factor claim 1 , comprise one or more of:a bit-depth of pixels of the respective portion, wherein the encoder generates or maintains different look-up tables for different bit-depths;a dynamic range associated with the respective portion, wherein the encoder generates or maintains different look-up tables for standard dynamic range and high-dynamic range;color primaries, wherein the encoder generates or maintains different look-up tables for different color primaries;transfer characteristics, wherein the encoder generates or maintains different look-up tables for different transfer characteristics;matrix coefficients, wherein the encoder generates or maintains different look-up tables for different matrix coefficients; orprediction mode type, wherein the encoder generates or maintains different look-up tables for an intra-prediction mode and an inter-prediction mode.4. The system of claim 1 , wherein the one or more characteristics of the respective portion used to determine the adaptive adjustment factor claim 1 , comprise one ...

Подробнее
03-03-2022 дата публикации

Point Cloud Compression with Multi-Resolution Video Encoding

Номер: US20220070493A1
Принадлежит: Apple Inc.

A system comprises an encoder configured to compress attribute and/or spatial information for a point cloud and/or a decoder configured to decompress compressed attribute and/or spatial information for the point cloud. To compress the attribute and/or spatial information, the encoder is configured to convert a point cloud into an image based representation. Also, the decoder is configured to generate a decompressed point cloud based on an image based representation of a point cloud. In some embodiments, an encoder performs downscaling of an image frame prior to video encoding and a decoder performs upscaling of an image frame subsequent to video decoding.

Подробнее
13-02-2020 дата публикации

POINT CLOUD GEOMETRY COMPRESSION

Номер: US20200053391A1
Принадлежит: Apple Inc.

A system comprises an encoder configured to compress a point cloud comprising a plurality of points each point comprising spatial information for the point. The encoder is configured to sub-sample the points and determine subdivision locations for the subsampled points. Also, the encoder is configured to determine, for respective subdivision location, if a point is to be included, not included, or relocated relative to the subdivision location. The encoder encodes spatial information for the sub-sampled points and encodes subdivision location point inclusion/relocation information to generate a compressed point cloud. A decoder recreates an original or near replica of an original point cloud based on the spatial information and the subdivision location inclusion/relocation information included in the compressed point cloud. 120-. (canceled)21. A non-transitory computer-readable medium storing program instructions that , when executed by one or more processors , cause the one or more processors to:sub-sample a point cloud, wherein the sub-sampled point cloud comprises fewer points than an original version of the point cloud; predict a predicted point location between the respective point of the sub-sampled point cloud and a neighboring point in the sub-sampled point cloud;', 'determine, based on comparing the predicted point location to the original version of the point cloud, update information for the predicted point location; and, 'for respective ones of the points of the sub-sampled point cloudencode data for a compressed version of the point cloud, the data comprising the determined update information for the predicted point locations.22. The non-transitory computer-readable medium of claim 21 , wherein the update information for a respective one of the predicted point locations indicates:whether a respective point is to be included at the predicted point location, not included at the predicted point location, or relocated relative to the predicted point ...

Подробнее
03-03-2016 дата публикации

Overlapped Rate Control for Video Splicing Applications

Номер: US20160065957A1

Rate control techniques are provided for encoding an input video sequence into a compressed coded bitstream with multiple coding passes. The final coding pass may comprise final splices with non-overlapping frames that do not extend into neighboring final splices. A final splice in the final coding pass may correspond to at least one non-final splice in a non-final coding pass. A non-final splice may have overlapping frames that extend into neighboring final splices in the final coding pass. The overlapping frames in the non-final splice may be used to derive complexity information about the neighboring final splices. The complexity information about the neighboring final splices, as derived from the overlapping frames, may be used to allocate or improve rate control related budgets in encoding the final splice into the compressed coded bitstream in the final coding pass. 1at least one non-transitory storage medium;a first compressed segment for a first period of a video sequence, the first compressed segment compressed at a first quality and stored by the at least one non-transitory storage medium;a second compressed segment for the second period of a video sequence, the second compressed segment compressed at a second quality and stored by the at least one non-transitory storage medium;a third compressed segment for a second period of the video sequence stored by the at least one non-transitory storage medium,a two-way communication interface configured to stream a compressed video bitstream and receive instructions over a network;wherein:(i) the first and second periods are adjacent and non-overlapping;(ii) the first and second qualities differ;(iii) the first and third segments are coded independently; and(iv) the compressed video bitstream includes the third compressed segment and either the first or second compressed segments.. A video system comprising: This Application is a continuation application of U.S. patent application Ser. No. 13/784,677 filed on Mar. ...

Подробнее
01-03-2018 дата публикации

MULTIDIMENSIONAL QUANTIZATION TECHNIQUES FOR VIDEO CODING/DECODING SYSTEMS

Номер: US20180063544A1
Принадлежит:

Video compression and decompression techniques are disclosed that provide improved bandwidth control for video compression and decompression systems. In particular, video coding and decoding techniques quantize input video in multiple dimensions. According to these techniques, pixel residuals may be generated from a comparison of an array of input data to an array of prediction data. The pixel residuals may be quantized in a first dimension. After the quantization, the quantized pixel residuals may be transformed to an array of transform coefficients. The transform coefficients may be quantized in a second dimension and entropy coded. Decoding techniques invert these processes. In still other embodiments, multiple quantizers may be provided upstream of the transform stage, either in parallel or in cascade, which provide greater flexibility to video coders to quantize data in different dimensions in an effort to balance the competing interest in compression efficiency and quality of reconstructed video. 1. A video coding method , comprising:generating pixel residuals from a comparison of an array of input data to an array of prediction data,first quantizing the pixel residuals in a first dimension,transforming the quantized pixel residuals to an array of transform coefficients,second quantizing the transform coefficients in a second dimension, andentropy coding the quantized transform coefficients.2. The video coding method of claim 1 , wherein the first quantizing comprises downscaling an array of the pixel residuals.3. The video coding method of claim 1 , wherein the first quantizing comprises quantizing an array of the pixel residuals uniformly by color component.4. The video coding method of claim 1 , wherein the first quantizing comprises quantizing an array of the pixel residuals non-uniformly by color component.5. The video coding method of claim 1 , wherein the first quantizing comprises quantizing values of the pixel residuals.6. The video coding method of ...

Подробнее
12-03-2015 дата публикации

CHROMA QUANTIZATION IN VIDEO CODING

Номер: US20150071344A1
Принадлежит:

A method of signaling additional chroma QP offset values that are specific to quantization groups is provided, in which each quantization group explicitly specifies its own set of chroma QP offset values. Alternatively, a table of possible sets of chroma QP offset values is specified in the header area of the picture, and each quantization group uses an index to select an entry from the table for determining its own set of chroma QP offset values. The quantization group specific chroma QP offset values are then used to determine the chroma QP values for blocks within the quantization group in addition to chroma QP offset values already specified for higher levels of the video coding hierarchy. 1. A method comprising:identifying one or more initial sets of chroma quantization parameter (QP) offset values at one or more levels of a video coding hierarchy, each set of chroma QP offset values at a particular level for specifying chroma QPs of video units encompassed by said particular level;identifying an additional set of chroma QP offset values for a plurality of video units in the video coding hierarchy; andcomputing a set of chroma QP values for the plurality of video units, wherein the identified initial sets of chroma QP offset values and the identified additional set of chroma QP offset values are additive components of the set of chroma QP value for the plurality of video units.2. The method of claim 1 , wherein the identified initial sets of chroma quantization parameters are for video units that encompass the particular level of the video coding hierarchy.3. The method of further comprising identifying a luma quantization parameter value for the plurality of video units.4. The method of claim 3 , wherein computing the set of chroma QP values comprises adding the identified initial sets of chroma QP offset values and the identified additional set of chroma QP offset values to a luma QP value.5. The method of claim 3 , wherein identifying the additional set of ...

Подробнее
12-03-2015 дата публикации

CHROMA QUANTIZATION IN VIDEO CODING

Номер: US20150071345A1
Принадлежит:

A method of signaling additional chroma QP offset values that are specific to quantization groups is provided, in which each quantization group explicitly specifies its own set of chroma QP offset values. Alternatively, a table of possible sets of chroma QP offset values is specified in the header area of the picture, and each quantization group uses an index to select an entry from the table for determining its own set of chroma QP offset values. The quantization group specific chroma QP offset values are then used to determine the chroma QP values for blocks within the quantization group in addition to chroma QP offset values already specified for higher levels of the video coding hierarchy. 1. A method comprising:receiving an array of sets of chroma quantization parameter (QP) offset values from an header of an encoded video picture, wherein an encoded video picture includes a plurality of quantization groups, each quantization group includes at least one coding unit;selecting one set of chroma QP offset values form the array of sets of chroma QP offset values for a particular quantization group from the plurality of quantization groups; andcomputing a set of chroma quantization parameters for a coding unit in the particular quantization group based on the selected set of chroma QP offset values.2. The method of claim 1 , wherein computing the set of chroma quantization parameters comprises adding the selected set of chroma QP offset values to a luma QP value.3. The method of further comprising identifying the luma QP value for the coding unit.4. The method of claim 2 , wherein the selected set of chroma QP offset values is a first set of chroma QP offset values claim 2 , the method further comprising adding a second set of chroma QP offset values associated with the video picture to the luma QP value and the first set of chroma QP offset values.5. The method of claim 2 , wherein the selected set of chroma QP offset values is a first set of chroma QP offset ...

Подробнее
10-03-2016 дата публикации

Compatible Stereoscopic Video Delivery

Номер: US20160073084A1

Stereoscopic images are subsampled and placed in a “checkerboard” pattern in an image. The image is encoded in a monoscopic video format. The monoscopic video is transmitted to a device where the “checkerboard” is decoded. Portions of the checkerboard (e.g., “black” portions) are used to reconstruct one of the stereoscopic images and the other portion of the checkerboard (e.g., “white” portions) are used to reconstruct the other image. The subsamples are, for example, taken from the image in a location coincident to the checkerboard position in which the subsamples are encoded. 1. A method of decoding with a processor a multiplexed image in a sequence of images , the method comprising:receiving a coded bitstream comprising a coded image in a multiplexed format, the coded image comprising samples of a first image and samples of a second image, wherein the multiplexing format is determined using SEI messaging in the coded bitstream;decoding with the processor the coded image to generate a decoded image;demultiplexing with the processor the decoded image to generate decoded samples of the first image and decoded samples of the second image, wherein the decoded samples of the first image and the second image are arranged in alternating order within even lines of the decoded image and in the opposite order in odd lines of the decoded image, wherein the decoded samples of the first image represent samples of the first image in a first checkerboard pattern and the decoded samples of the second image represent samples of the second image in a second checkerboard pattern;upsampling with the processor the decoded samples of the first image to generate a first output image; andupsampling with the processor the decoded samples of the second image to generate a second output image.2. The method of claim 1 , wherein sampling an image with even and odd pixel columns in a first checkerboard pattern comprises:on even rows, sampling only the even columns of the image; andon odd rows ...

Подробнее
08-03-2018 дата публикации

Compatible stereoscopic video delivery

Номер: US20180070071A1
Принадлежит: Dolby Laboratories Licensing Corp

Stereoscopic images are subsampled and placed in a “checkerboard” pattern in an image. The image is encoded in a monoscopic video format. The monoscopic video is transmitted to a device where the “checkerboard” is decoded. Portions of the checkerboard (e.g., “black” portions) are used to reconstruct one of the stereoscopic images and the other portion of the checkerboard (e.g., “white” portions) are used to reconstruct the other image. The subsamples are, for example, taken from the image in a location coincident to the checkerboard position in which the subsamples are encoded.

Подробнее
28-02-2019 дата публикации

SYSTEMS AND METHODS FOR MULTI-LAYERED FRAME-COMPATIBLE VIDEO DELIVERY

Номер: US20190068947A1

Multi-layered frame-compatible video delivery is described. Multi-layered encoding and decoding methods, comprising a base layer and at least one enhancement layer with reference processing, are provided. In addition, multi-layered encoding and decoding methods with inter-layer dependencies are described. Encoding and decoding methods that are capable of frame-compatible 3D video delivery are also described. 1. A decoding method for multi-layered frame compatible video delivery , the method comprising: receiving an input video signal comprising input frames encoded in a multi-layer format, wherein the multi-layer format comprises:', 'a coded base layer frame comprising even rows from a first view and odd rows from a second view interleaved in a top-bottom packing arrangement;', 'a coded enhancement layer frame comprising odd rows from the first view and even rows from the second view interleaved in top-bottom packing arrangement;, 'in a decoder comprising one or more processors,'}decoding and unpacking the coded base layer frame to generate a decoded base layer first view and a decoded base layer second view;decoding and unpacking the coded enhancement layer (EL) frame to generate a decoded enhancement layer first view and a decoded enhancement layer second view; andupsampling the decoded base layer first view or the decoded base layer second view to generate an output frame in a frame compatible format.2. The method of claim 1 , further comprising:combining the decoded base layer first view and the decoded enhancement layer first view to generate a first 3D view; andcombining the decoded base layer second view and the decoded enhancement layer second view to generate a second 3D view.3. The method of claim 1 , wherein the decoding and unpacking of the enhancement layer frame uses a first reference picture buffer for generating the decoded enhancement layer first view and a second reference picture buffer for generating the decoded enhancement layer second view.4. ...

Подробнее
28-02-2019 дата публикации

TECHNIQUES FOR RESOURCE CONSERVATION DURING PERFORMANCE OF INTRA BLOCK COPY PREDICTION SEARCHES

Номер: US20190068992A1
Принадлежит:

Methods are described for encoding and decoding blocks of image data using intra block copying (IBC). A source block for intra block copying is selected from a source region of a current image that is closer to the current block than a threshold, wherein the source region does not include a portion of the current image that is further from the current block than the threshold. 122-. (canceled)23. A system for video decoding , comprising a processor and memory , the memory containing instructions that , when executed on the processor , cause the system to at least:decode coded video data received from a channel on a pixel block-by-pixel block basis, wherein the coded video data includes, for a given pixel block, an indication of a reference pixel block and wherein the decoding comprises predicting the given pixel block from within a source region of a buffer;store the decoded pixel blocks in the buffer;determine a height limit for limiting the source region measured in samples of video data;following the decoding of the given pixel block, adjust the source region to include a first portion of a current image that is closer to the given pixel block than the height limit and to not include a second portion of the current image that is further from the current block than the height limit.24. The system of claim 23 , wherein the current image is divided into tiles claim 23 , the given pixel block and the source region are both contained with a current tile claim 23 , and the height limit is measured within the current tile.25. The system of claim 23 , wherein the instructions further cause the system to:in-loop filter the source region of the buffer prior to the predicting the given block.26. The system of claim 23 , wherein the instructions further cause the system to:in-loop filter the source region of the buffer after the predicting the given block.27. The system of claim 23 , wherein the height limit is measured in pixel rows.28. The system of claim 23 , wherein the ...

Подробнее
19-03-2015 дата публикации

Device and Method for Fast Block-Matching Motion Estimation in Video Encoders

Номер: US20150078451A1
Принадлежит:

A solution is provided to estimate motion vectors of a video. A multistage motion vector prediction engine is configured to estimate multiple best block-matching motion vectors for each block in each video frame of the video. For each stage of the motion vector estimation for a block of a video frame, the prediction engine selects a test vector form a predictor set of test vectors, computes a rate-distortion optimization (RDO) based metric for the selected test vector, and selects a subset of test vectors as individual best matched motion vectors based on the RDO based metric. The selected individual best matched motion vectors are compared and a total best matched motion vector is selected based on the comparison. The prediction engine selects iteratively applies one or more global matching criteria to the selected best matched motion vector to select a best matched motion vector for the block of pixels. 1. An apparatus for estimating motion vectors of a video , comprising:a computer readable medium storing the video comprising a plurality of video frames, each frame comprising a plurality of blocks; and selecting a test vector from a predictor set of test vectors, wherein the set of test vectors is selected from a plurality of predictor sets of vectors based on using at least one of a priori knowledge of the video and a priori knowledge of a plurality of video sequences stored in a database, and wherein each test vector is unique from other test vectors of the predictor set of the test vectors;', 'computing a rate-distortion optimization (RDO) based metric for the selected test vector;', 'selecting a subset of test vectors as individual best matched motion vectors based on the RDO based metric;', 'selecting a total best matched motion vector from the individual best matched motion vectors based on comparisons of the individual best matched motion vectors; and', 'iteratively applying one or more global matching criteria to the selected total best matched motion ...

Подробнее
11-03-2021 дата публикации

TEMPORALLY-OVERLAPPED VIDEO ENCODING, VIDEO DECODING AND VIDEO RENDERING TECHNIQUES THEREFOR

Номер: US20210076054A1
Принадлежит:

Techniques for video exchange are disclosed in which a video source sends redundant copies of video coded at a given presentation time. A decoder may determine, from metadata stored with a first decoded frame, whether other decoded frames have been designated as correlated to the first decoded frame. If so, the decoder may fetch other decoding frames using time indicator values of other decoded frames in metadata stored with the first decoded frame. When other decoded frame(s) are found, the decoder may blend content of the first coded frame and the other decoded frame(s). The decoder may render the blended frame. When one such coded frame is an intra-coded frame, the proposed technique may alleviate beating artifacts by blending recovered data of the intra-coded frame with recovered data of another frame, which may be inter-coded. 1. A method , comprising:determining, from metadata stored with a first decoded frame from a first coded video sequence, whether another decoded frame is having a common temporally correlated to instance with the first decoded frame;when another decoded frame is having a common temporal instance with the first decoded frame, blending content of the first decoded frame and the other decoded frame; andrendering the blended frame.20. The method of claim , wherein the first decoded frame is a decoded intra-coded frame.30. The method of claim , wherein separate coded video sequences are demarcated by respective intra random access point frames.40. The method of claim , wherein the coded video sequences are respective Groups of Frames.50. The method of claim , wherein the coded video sequences are respective segments identified on a manifest file.60. The method of claim , further comprising , when no other decoded frames are designated by the metadata of the first decoded frame , rendering the first decoded frame.70. The method of claim , wherein , when a third decoded frame is temporally correlated to the first decoded frame , the blending ...

Подробнее
07-03-2019 дата публикации

MACHINE LEARNING VIDEO PROCESSING SYSTEMS AND METHODS

Номер: US20190075301A1
Принадлежит:

System and method for improving video encoding and/or video decoding. In embodiments, a video encoding pipeline includes a main encoding pipeline that compresses source image data corresponding with an image frame by processing the source image data based at least in part on encoding parameters to generate encoded image data. Additionally the video encoding pipeline includes a machine learning block communicatively coupled to the main encoding pipeline, in which the machine learning block analyzes content of the image frame by processing the source image data based at least in part on machine learning parameters implemented in the machine learning block when the machine learning block is enabled by the encoding parameters; and the video encoding pipeline adaptively adjusts the encoding parameters based at least in part on the content expected to be present in the image frame to facilitate improving encoding efficiency. 1. A video encoding pipeline comprising:a main encoding pipeline implemented in an electronic device, wherein the main encoding pipeline is configured to compress source image data corresponding with at least a portion of an image frame by processing the source image data based at least in part on encoding parameters to generate encoded image data; and the machine learning block is configured to analyze content of the image frame by processing the source image data based at least in part on machine learning parameters implemented in the machine learning block when the machine learning block is enabled by the encoding parameters; and', 'the video encoding pipeline is configured to adaptively adjust the encoding parameters based at least in part on the content expected to be present in the image frame to facilitate improving encoding efficiency., 'a machine learning block communicatively coupled to the main encoding pipeline, wherein2. The video encoding pipeline of claim 1 , wherein:the machine learning block comprises a convolutional neural network ...

Подробнее
07-03-2019 дата публикации

POINT CLOUD GEOMETRY COMPRESSION

Номер: US20190075320A1
Принадлежит: Apple Inc.

A system comprises an encoder configured to compress a point cloud comprising a plurality of points each point comprising spatial information for the point. The encoder is configured to sub-sample the points and determine subdivision locations for the subsampled points. Also, the encoder is configured to determine, for respective subdivision location, if a point is to be included, not included, or relocated relative to the subdivision location. The encoder encodes spatial information for the sub-sampled points and encodes subdivision location point inclusion/relocation information to generate a compressed point cloud. A decoder recreates an original or near replica of an original point cloud based on the spatial information and the subdivision location inclusion/relocation information included in the compressed point cloud. 1. A system comprising:one or more sensors configured to capture a plurality of points that make up a point cloud, wherein each of the respective points comprises spatial information identifying a spatial location of the respective point; and sub-sample a captured point cloud captured by the one or more sensors, wherein the sub-sampled point cloud comprises fewer points than the captured point cloud; and', identify a location between the respective point of the sub-sampled point cloud and a neighboring point in the sub-sampled point cloud; and', 'determine, based on comparing the location to the captured point cloud, whether a point in a decompressed point cloud is to be included at the location, not included at the location, or relocated relative to the location; and, 'for respective ones of the points of the sub-sampled point cloud, spatial information for the points of the sub-sampled point cloud; and', 'data indicating, for each of the respective locations, whether a respective point is to be included at the location, not included at the location, or relocated relative to the location in the decompressed point cloud., 'encode data for the ...

Подробнее
17-03-2016 дата публикации

Encoding and Decoding Architecture of Checkerboard Multiplexed Image Data

Номер: US20160080772A1

A device includes a coder or a codec configured for interleaved image data utilizing diamond shaped blocks for motion estimation and/or motion compensation and utilizing square or orthogonal transforms of residual data. In various embodiments, the decoder may be configured, among others, to perform de-blocking on edges of the diamond shaped blocks and/or data padding at boundaries of the image data. 1. A method of decoding interleaved image data including the step of utilizing diamond shaped blocks for motion estimation and/or motion compensation and utilizing square or orthogonal transforms of residual data.2. The method according to claim 1 , wherein de-blocking is performed on edges of the diamond shaped blocks.3. The method according to claim 1 , wherein data padding is performed at boundaries of the image data.4. The method according to claim 1 , wherein:said method is embodied in a set of computer instructions stored on a computer readable media;said computer instructions, when loaded into a computer, cause the computer to perform the steps of said method.5. The method according to claim 4 , wherein said computer instruction are compiled computer instructions stored as an executable program on said computer readable media.6. A device claim 4 , comprising a codec configured for interleaved image data utilizing diamond shaped blocks for motion estimation and/or motion compensation and utilizing square or orthogonal transforms of residual data.7. The device according to claim 6 , wherein the codec is further configured to perform at least one of de-blocking performed on edges of the diamond shaped blocks and data padding at boundaries of the image data.8. The device according to claim 6 , wherein the device comprises at least one of a set-top box claim 6 , DVD player claim 6 , Blu-ray players claim 6 , and a 3D capable display.9. The device according to claim 6 , wherein the device comprises a playback device and the interleaved image data is interleaved left and ...

Подробнее
18-03-2021 дата публикации

Point Cloud Compression Using Fixed-Point Numbers

Номер: US20210082152A1
Принадлежит: Apple Inc.

A system comprises an encoder configured to compress attribute information for a point cloud and/or a decoder configured to decompress compressed attribute information. Attribute values for at least one starting point are included in a compressed attribute information file and attribute correction values are included in the compressed attribute information file. Attribute values are predicted based, at least in part, on attribute values of neighboring points. The predicted attribute values are compared to attribute values of a point cloud prior to compression to determine attribute correction values. In order to improve computing efficiency and/or repeatability, fixed-point number representations are used when determining predicted attribute values and attribute correction values. A decoder follows a similar prediction process as an encoder and corrects predicted values using attribute correction values included in a compressed attribute information file using fixed-point number representations. 1. A non-transitory computer-readable medium storing program instructions that , when executed by one or more processors , cause the one or more processors to: identify a set of neighboring points that neighbor the respective point;', the one or more attribute values are predicted using an inverse-distance based interpolation that uses attribute values of the set of neighboring points and distances to the set of neighboring points to predict the one or more attribute values of the respective point;', 'the inverse-distance based interpolation is performed using fixed-point number representations; and', 'a division operation of the inverse distance-based interpolation is performed using a look-up table comprising approximated division results represented in fixed-point number representations; and, 'predict one or more attribute values for the respective point, wherein, 'determine one or more attribute correction values for the respective point based on comparing the one or ...

Подробнее
14-03-2019 дата публикации

HIERARCHICAL POINT CLOUD COMPRESSION

Номер: US20190081638A1
Принадлежит: Apple Inc.

A system comprises an encoder configured to compress attribute information for a point cloud and/or a decoder configured to decompress compressed attribute information for the point cloud. Attribute values for at least one starting point are included in a compressed attribute information file and attribute correction values used to correct predicted attribute values are included in the compressed attribute information file. Attribute values are predicted based, at least in part, on attribute values of neighboring points and distances between a particular point for whom an attribute value is being predicted and the neighboring points. The predicted attribute values are compared to attribute values of a point cloud prior to compression to determine attribute correction values. A decoder follows a similar prediction process as an encoder and corrects predicted values using attribute correction values included in a compressed attribute information file. 1. A system comprising:one or more sensors configured to capture a plurality of points that make up a point cloud, wherein respective ones of the points comprise spatial information for the point and attribute information for the point; and determine a first level of detail for the attribute information of the point cloud; and', 'determine one or more additional levels of detail for the attribute information of the point cloud,, 'an encoder configured to assign an attribute value to at least one point of the point cloud based on the attribute information included in the captured point cloud for the point; and', identify a set of neighboring points greater than a first distance from the point;', 'determine a predicted attribute value for the respective point based on predicted or assigned attributes values for the neighboring points; and', 'determine, based on comparing the predicted attribute value for the respective point to the attribute information for the point included in the captured point cloud, an attribute ...

Подробнее
22-03-2018 дата публикации

PREDICTIVE MOTION VECTOR CODING

Номер: US20180084255A1
Принадлежит:

Overlapped block disparity estimation and compensation is described. Compensating for images with overlapped block disparity compensation (OBDC) involves determining if OBDC is enabled in a video bit stream, and determining if OBDC is enabled for one or more macroblocks that neighbor a first macroblock within the video bit stream. The neighboring macroblocks may be transform coded. If OBDC is enabled in the video bit stream and for the one or more neighboring macroblocks, predictions may be made for a region of the first macroblock that has an edge adjacent with the neighboring macroblocks. OBDC can be causally applied. Disparity compensation parameters or modes may be shared amongst views or layers. A variety of predictions may be used with causally-applied OBDC. 1. (canceled)2. A method to encode for a block motion compensated video bit stream , the method comprising:encoding, by an encoder, a first block partition having a prediction type of inter-prediction,encoding a second block partition having a prediction type of inter-prediction, the second block partition being edge adjacent to the first block partition;encoding a third block partition; andencoding an explicit signal, the explicit signal indicating a mode for the first block partition,wherein a first mode identified by the explicit signal identifies the first block partition as being block motion compensated by motion vector information of the second block partition,wherein a second mode identified by the explicit signal identifies the first block partition as being block motion compensated by motion vector information of the third block partition without motion vector information of the second block partition, andwherein the encoder includes a processor configured to perform instructions.3. A method to encode a motion compensated video bit stream by a mobile telephone , the method comprising:encoding, by the mobile telephone, a first block partition having a prediction type of inter-prediction,encoding a ...

Подробнее
22-03-2018 дата публикации

CODING AND DECODING OF INTERLEAVED IMAGE DATA

Номер: US20180084278A1
Принадлежит:

Sampled data is packaged in checkerboard format for encoding and decoding. The sampled data may be quincunx sampled multi-image video data (e.g., 3D video or a multi-program stream), and the data may also be divided into sub-images of each image which are then multiplexed, or interleaved, in frames of a video stream to be encoded and then decoded using a standardized video encoder. A system for viewing may utilize a standard video decoder and a formatting device that de-interleaves the decoded sub-images of each frame reformats the images for a display device. A 3D video may be encoded using a most advantageous interleaving format such that a preferred quality and compression ratio is reached. In one embodiment, the invention includes a display device that accepts data in multiple formats. 1. (canceled)2. A video encoding method , comprising:sub-sampling a first image and a second image into a first sampled image data and a second sampled image data respectively via horizontal sampling;separating each of the first sampled image data and the second sampled image data into a plurality of sub-images, wherein each of the sub-images comprises image data of the corresponding sampled image data selected via a pattern;packaging at least the plurality of sub-images together into a single image frame of a video stream, wherein the packaging comprises tiling the sub-images together according to an arrangement indicating how the sub-images are tiled together, and wherein the arrangement indicates that the plurality of sub-images corresponding to the first sampled image data are tiled into a first rectangular block of the image frame, and the plurality of sub-images corresponding to the second sampled image data are tiled into a second rectangular block of the image frame, and wherein the first rectangular block is positioned in side-by-side relation with the second rectangular block in the image frame; andencoding the single image frame, and encoding an identifier of the ...

Подробнее
31-03-2022 дата публикации

MULTIDIMENSIONAL QUANTIZATION TECHNIQUES FOR VIDEO CODING/DECODING SYSTEMS

Номер: US20220103851A1
Принадлежит:

Video compression and decompression techniques are disclosed that provide improved bandwidth control for video compression and decompression systems. In particular, video coding and decoding techniques quantize input video in multiple dimensions. According to these techniques, pixel residuals may be generated from a comparison of an array of input data to an array of prediction data. The pixel residuals may be quantized in a first dimension. After the quantization, the quantized pixel residuals may be transformed to an array of transform coefficients. The transform coefficients may be quantized in a second dimension and entropy coded. Decoding techniques invert these processes. In still other embodiments, multiple quantizers may be provided upstream of the transform stage, either in parallel or in cascade, which provide greater flexibility to video coders to quantize data in different dimensions in an effort to balance the competing interest in compression efficiency and quality of reconstructed video. 129.-. (canceled)30. A video decoding method , comprising:decoding reference image data of a coded video sequence of frames received from a channel;predicting blocks of prediction data from the decoded reference image data;entropy decoding data of the coded video sequence, yielding blocks of residual transform coefficients and corresponding quantization parameters including an indication of residual downsampling;transforming the residual transform coefficients in a transform-domain into blocks of reconstructed residual data in a pixel-domain;when the quantization parameters indicate residual downsampling was applied only to chroma components of a first block, upscaling the chroma components of the corresponding reconstructed residual data, and generating a block of reconstructed pixel values from the corresponding block of the prediction data, the corresponding block of the upscaled chroma components of reconstructed residual data, and the corresponding block the ...

Подробнее
25-03-2021 дата публикации

Three-Dimensional Mesh Compression Using a Video Encoder

Номер: US20210090301A1
Принадлежит: Apple Inc.

A system comprises an encoder configured to compress and encode data for a three-dimensional mesh using a video encoding technique. To compress the three-dimensional mesh, the encoder determines sub-meshes and for each sub-mesh: texture patches and geometry patches. Also the encoder determines patch connectivity information and patch texture coordinates for the texture patches and geometry patches. The texture patches and geometry patches are packed into video image frames and encoded using a video codec. Additionally, the encoder determines boundary stitching information for the sub-meshes. A decoder receives a bit stream as generated by the encoder and reconstructs the three-dimensional mesh. 1. A non-transitory computer-readable medium storing program instructions that , when executed by one or more processors , cause the one or more processors to: one or more attribute or texture patches; and', 'for each of a plurality of sub-meshes of the three-dimensional mesh, a geometry patch that indicates depth values for vertices of the respective sub-mesh;, 'determine, for a three-dimensional mesh, a plurality of patches, wherein the patches comprisepack two-dimensional patch images for the determined patches into a plurality of two-dimensional image frames;determine patch connectivity information and patch texture or attribute coordinates for the patches;determine boundary stitching information to be used to join the sub-meshes back into the three-dimensional mesh; andencode the determined patches, the determined patch connectivity and patch texture or attribute coordinates, and the determined boundary stitching information,wherein the determined patches are encoded via a video-based encoder as packed two-dimensional patch images packed into the plurality of two-dimensional image frames.2. The non-transitory computer-readable medium of claim 1 , wherein the three-dimensional mesh represents a dynamic object claim 1 , and wherein the program instructions claim 1 , when ...

Подробнее
21-03-2019 дата публикации

POINT CLOUD COMPRESSION USING NON-CUBIC PROJECTIONS AND MASKS

Номер: US20190087978A1
Принадлежит: Apple Inc.

A system comprises an encoder configured to compress attribute information and/or spatial for a point cloud and/or a decoder configured to decompress compressed attribute and/or spatial information for the point cloud. To compress the attribute and/or spatial information, the encoder is configured to convert a point cloud into an image based representation. Also, the decoder is configured to generate a decompressed point cloud based on an image based representation of a point cloud. 1. A system comprising:one or more sensors configured to capture a plurality of points that make up a point cloud, wherein respective ones of the points comprise spatial information for the point and attribute information for the point; and generate, for the point cloud, one or more projected images;', 'pack the generated one or more projected images into one or more image frames;', 'pad spaces in the one or more image frames that are not covered by the one or more projected images with a pad;', 'encode the one or more image frames; and', 'encode information indicating regions of the one or more image frames that correspond to active regions or non-active regions of the one or more image frames, wherein regions that are covered by the pad are indicated as non-active regions., 'an encoder configured to compress the point cloud, wherein to compress the point cloud, the encoder is configured to2. The system of claim 1 , wherein to generate the one or more projected images claim 1 , the encoder is configured to:project the point cloud onto a cylindrical surface or a spherical surface.3. The system of claim 1 , wherein to generate the one or more projected images claim 1 , the encoder is configured to:project portions of the point cloud onto three or more parallel planes.4. The system of claim 1 , wherein the encoder is further configured to:encode, for points of the point cloud that are not visible from a particular point-of-view, information indicating the points are non-active points in ...

Подробнее
21-03-2019 дата публикации

POINT CLOUD COMPRESSION

Номер: US20190087979A1
Принадлежит: Apple Inc.

A system comprises an encoder configured to compress attribute information and/or spatial for a point cloud and/or a decoder configured to decompress compressed attribute and/or spatial information for the point cloud. To compress the attribute and/or spatial information, the encoder is configured to convert a point cloud into an image based representation. Also, the decoder is configured to generate a decompressed point cloud based on an image based representation of a point cloud. 1. A system comprising:one or more sensors configured to capture a plurality of points that make up a point cloud, wherein respective ones of the points comprise spatial information for the point and attribute information for the point; and determine, for the point cloud, a plurality of patches each corresponding to portions of the point cloud, wherein each patch comprises points with surface normal vectors that deviate from one another less than a threshold amount;', generate a patch image comprising the set of points corresponding to the patch projected onto a patch plane; and', 'generate a patch image comprising depth information for the set of points corresponding to the patch, wherein the depth information represents depths of the points in a direction perpendicular to the patch plane;, 'for each patch,'}, 'pack generated patch images for each of the determined patches into one or more image frames; and, 'an encoder configured to compress the point cloud, wherein to compress the point cloud, the encoder is configured toencode the one or more image frames.2. The system of claim 1 , wherein the encoder is further configured to claim 1 , for each patch:generate a patch image comprising attribute information for one or more attributes of the set of points corresponding to the patch;wherein said pack generated patch images for the determined patches into one or more image frames comprises packing the patch image comprising attribute information.3. The system of claim 1 , wherein the ...

Подробнее
25-03-2021 дата публикации

Video Compression with In-Loop Sub-Image Level Controllable Noise Generation

Номер: US20210092459A1
Принадлежит:

A system comprises an encoder configured to compress video data using an in-loop noise generation process that generates noise in the compression loop at a sub-image portion level of granularity, such as at a block level. The encoder includes noise model and/or noise model input parameter information in an encoded bit stream. Also, a system includes a decoder configured to receive such a bit stream and decompress the video using an in-loop noise generation process that generates noise in the decompression loop at a sub-image portion level of granularity. 1. A system configured to encode video image data , the system comprising:one or more processors; and [ determine a first noise pattern to be generated for a first portion of a given one of the images;', 'determine a second noise pattern to be generated for a second portion of the same given one of the images;', 'generate, via an in-loop noise generator included in a compression loop for compressing the pixel value data, the first noise pattern and the second noise pattern, wherein the first and second noise patterns are different; and, 'compress pixel value data for a plurality of images included in a plurality of frames of a video, wherein to compress the pixel value data, the program instructions cause the one or more processors to, 'encode the compressed pixel value data for the plurality of images, wherein an encoded bit stream for the compressed pixel value data includes data indicating a noise model or one or more noise parameters used to generate the first noise pattern and the second noise pattern., 'one or more memories storing program instructions, that when executed on or across the one or more processors, cause the one or more processors to2. The system of claim 1 , wherein the first portion of the given image corresponds to a block of the given image and the second portion corresponds to another bock of the given image claim 1 ,wherein the program instructions, when executed on or across the one or ...

Подробнее
31-03-2016 дата публикации

Directed Interpolation and Data Post-Processing

Номер: US20160094859A1

An encoding device evaluates a plurality of processing and/or post-processing algorithms and/or methods to be applied to a video stream, and signals a selected method, algorithm, class or category of methods/algorithms either in an encoded bitstream or as side information related to the encoded bitstream. A decoding device or post-processor utilizes the signaled algorithm or selects an algorithm/method based on the signaled method or algorithm. The selection is based, for example, on availability of the algorithm/method at the decoder/post-processor and/or cost of implementation. The video stream may comprise, for example, downsampled multiplexed stereoscopic images and the selected algorithm may include any of upconversion and/or error correction techniques that contribute to a restoration of the downsampled images. 1. A method for decoding at least one temporal cuboid region of a plurality of pictures , the method comprising:decoding, by a decoder, a video bitstream comprising a plurality of pictures with two interleaved images, the two interleaved images being quincunx sampled and checkerboard multiplexed;de-interleaving and upsampling the two images;after de-interleaving and upsampling, storing the two images in a buffer; andreceiving, by the decoder, metadata from the video bitstream indicating a width, height, and temporal length of the at least one temporal cuboid region for the plurality of pictures, the width and height being fixed for the temporal length,wherein the at least one cuboid region is smaller in a width or height than a picture of the plurality of pictures,the at least one temporal cuboid region is classified by semantic information, andthe temporal length is greater than one picture.2. The method of wherein the at least one cuboid region is two regions.3. The method of wherein the two regions overlap.4. The method of wherein the two regions are non-overlapping.5. The method of wherein the plurality of pictures is a Group of Pictures.6. The ...

Подробнее
07-04-2016 дата публикации

CHROMA QUANTIZATION IN VIDEO CODING

Номер: US20160100170A1
Принадлежит:

A method of signaling additional chroma QP offset values that are specific to quantization groups is provided, in which each quantization group explicitly specifies its own set of chroma QP offset values. Alternatively, a table of possible sets of chroma QP offset values is specified in the header area of the picture, and each quantization group uses an index to select an entry from the table for determining its own set of chroma QP offset values. The quantization group specific chroma QP offset values are then used to determine the chroma QP values for blocks within the quantization group in addition to chroma QP offset values already specified for higher levels of the video coding hierarchy. 119-. (canceled)20. A method comprising:receiving a plurality of sets of chroma quantization parameter (QP) offset values associated with the encoded video picture, wherein each set of chroma QP offset values in the plurality of sets of chroma QP offset values is associated with an index;receiving an encoded video picture comprising a plurality of quantization groups, wherein each quantization group is associated with an index; andcomputing a set of chroma quantization parameters for the video picture by computing a set of chroma QP offset values for each quantization group of the video picture by using the index of the quantization group to select a set chroma QP offset value from the plurality of sets of chroma QP offset values.21. The method of claim 20 , wherein computing the set of chroma quantization parameters comprises adding the selected set of chroma QP offset values to a luma QP value.22. The method of claim 21 , wherein the selected set of chroma QP offset values is a first set of chroma QP offset values claim 21 , the method further comprising adding a second set of chroma QP offset values associated with the video picture to the luma QP value and the first set of chroma QP offset values.23. The method of claim 20 , wherein each video comprises a plurality of ...

Подробнее
01-04-2021 дата публикации

PREDICTING ATTRIBUTES FOR POINT CLOUD COMPRESSION ACCORDING TO A SPACE FILLING CURVE

Номер: US20210097722A1
Принадлежит: Apple Inc.

Attributes for point cloud compression may be predicted according to a space filling curve. An order for selecting points of a point cloud to be compressed in order to predict attributes of the points may be determined according to values of a space filling curve applied to spatial information for the points. A point of the point cloud may be selected according to the determined ordering. A prediction technique may be performed with respect to the selected point based on a set of neighboring points in the point cloud also selected according to the ordering. Predicted attribute values and correction values may be determined for the predicted values. The correction values may be encoded as part of a compressed version of the point cloud. 1. A non-transitory computer-readable medium storing program instructions that , when executed by one or more processors , cause the one or more processors to: determine an order for the points of the three-dimensional point cloud according to a space filling curve applied to spatial information of the points of the three-dimensional point cloud; and', determine a set of neighborhood points for the given point being evaluated based on the determined order for the points according to the space filling curve, wherein the set of neighborhood points comprises a number of points included in the order at preceding or subsequent positions in the order that precede or follow the given point being evaluated;', 'determine from the set of neighborhood points, a sub-set of the neighborhood points that are nearest in location in three-dimensional space to the given point being evaluated;', 'predict an attribute value for the given point being evaluated based on respective attribute values of points included in the sub-set of the neighborhood points that are nearest to the given point being evaluated; and', 'determine an attribute correction value for the given point being evaluated based on comparing the predicted attribute value for the given ...

Подробнее
01-04-2021 дата публикации

POINT CLOUD COMPRESSION USING VIDEO ENCODING WITH TIME CONSISTENT PATCHES

Номер: US20210097723A1
Принадлежит: Apple Inc.

A system comprises an encoder configured to compress attribute and/or spatial information for a point cloud and/or a decoder configured to decompress compressed attribute and/or spatial information for the point cloud. To compress the attribute and/or spatial information, the encoder is configured to convert a point cloud into an image based representation. Also, the decoder is configured to generate a decompressed point cloud based on an image based representation of a point cloud. In some embodiments, an encoder generates time-consistent patches for multiple version of the point cloud at multiple moments in time and uses the time-consistent patches to generate image based representations of the point cloud at the multiple moments in time. 1. A non-transitory computer-readable medium storing program instructions that , when executed by one or more processors , cause the one or more processors to: determine one or more motion functions that model motion of one or more segments of the point cloud across multiple moments in time;', 'determine, based on the one or more motion functions, mappings between points of the point cloud across the multiple moments in time; and', 'determine one or more patches comprising points of the point cloud that have surface normal vectors that deviate from one another at each of the multiple moments in time by less than a threshold amount; and, 'determine, for a set of visual volumetric content comprising a point cloud, a plurality of patches, wherein each patch comprises points of the point cloud projected on a patch plane, wherein to determine the plurality of patches the program instructions cause the one or more processors topack patch images for the determined one or more patches into a plurality of image frames, wherein the plurality of image frames comprise consistent patches for the point cloud across the multiple moments in time.2. The non-transitory computer-readable medium of claim 1 , wherein the program instructions cause ...

Подробнее
01-04-2021 дата публикации

Point Cloud Compression with Closed-Loop Color Conversion

Номер: US20210097725A1
Принадлежит: Apple Inc.

A system comprises an encoder configured to compress attribute information and/or spatial for a point cloud and/or a decoder configured to decompress compressed attribute and/or spatial information for the point cloud. To compress the attribute and/or spatial information, the encoder is configured to convert a point cloud into an image based representation. Also, the decoder is configured to generate a decompressed point cloud based on an image based representation of a point cloud. A closed-loop color conversion process is used to improve compression while taking into consideration distortion introduced throughout the point cloud compression process. 120.-. (canceled)21. A non-transitory computer-readable medium storing program instructions that , when executed by one or more processors , cause the one or more processors to:receive a bit stream comprising one or more video encoded image frames, wherein the one or more video encoded image frames comprise packed patch images, wherein the patch images correspond to patches of a three-dimensional (3D) spatial and attribute data representation of an object projected onto a patch plane, and wherein the patch images have been converted from an original color space to a different color space;video decode the one or more video encoded image frames;convert the patch images from the different color space back to the original color space;reconstruct the 3D spatial and attribute data representation of the object using the converted patch images, wherein the reconstructed representation of the object has colors represented in the original color space.22. The non-transitory computer-readable medium of claim 21 , wherein the 3D spatial and attribute data representation of the object is a three-dimensional (3D) mesh of the object.23. The non-transitory computer-readable medium of claim 21 , wherein the 3D spatial and attribute data representation of the object is a three-dimensional (3D) point cloud of the object.24. The non- ...

Подробнее
01-04-2021 дата публикации

Point Cloud Attribute Transfer Algorithm

Номер: US20210097726A1
Принадлежит: Apple Inc.

A system comprises an encoder configured to compress attribute information and/or spatial for a point cloud and/or a decoder configured to decompress compressed attribute and/or spatial information for the point cloud. A point cloud attribute transfer algorithm may be used to determine distortion between an original point cloud and a reconstructed point cloud. Additionally, the point cloud attribute transfer algorithm may be used to select attribute values for a reconstructed point cloud such that distortion between an original point cloud and a reconstructed version of the original point cloud is minimized. 1. A system comprising:one or more sensors configured to capture a plurality of points that make up a point cloud, wherein respective ones of the points comprise spatial information for the point and attribute information for the point; and encode spatial information for the point cloud;', 'generate a re-constructed version of the point cloud based on the encoded spatial information, wherein at least some of the spatial locations of points or presence of points differs between the point cloud prior to the spatial information being encoded and the re-constructed version of the point cloud;', 'generate re-constructed attribute values for the points of the re-constructed version of the point cloud;', 'determine respective differences between re-constructed attribute values for points in the re-constructed version of the point cloud and attribute values of corresponding points in the point cloud prior to encoding; and', 'select modified attribute values to be encoded for the attribute values such that a level of attribute value distortion between the point cloud prior to encoding and the re-constructed version of the point cloud is minimized; and', 'encode modified attribute value information to be used at a decoder to generate a decoder generated re-constructed version of the point cloud, wherein the attribute value information is based on the modified attribute ...

Подробнее
19-03-2020 дата публикации

SCALABILITY OF MULTI-DIRECTIONAL VIDEO STREAMING

Номер: US20200092571A1
Принадлежит:

Aspects of the present disclosure provide techniques for reducing latency and improving image quality of a viewport extracted from multi-directional video communications. According to such techniques, first streams of coded video data are received from a source. The first streams include coded data for each of a plurality of tiles representing a multi-directional video, where each tile corresponding to a predetermined spatial region of the multi-directional video, and at least one tile of the plurality of tiles in the first streams contains a current viewport location at a receiver. The techniques include decoding the first streams and displaying the tile containing the current viewport location. When the viewport location at the receiver changes to include a new tile of the plurality of tiles, retrieving and decoding first streams for the new tile, displaying the decoded content for the changed viewport location, and transmitting the changed viewport location to the source. 1. A video reception method , comprising:receiving, from a source, first streams of coded data for each of a plurality of tiles representing a multi-directional video including a first tile encoded at a first tier of quality and other tiles of the plurality of tiles coded as a base layer of a second tier of quality, wherein each tile corresponds to a predetermined spatial region of the multi-directional video and a current viewport location at a receiver includes at least the first tile;decoding the first tier of the first tile from the first streams;displaying the decoded first tier for the current viewport location; decoding the base layer of the second tier of the second tile from first streams,', 'transmitting an indication of the second tile to the source;, 'when the viewport location at the receiver changes to include a second tile of the other tilesreceiving, from the source and in response to the transmitting, second streams of coded video data including the second tile coded as an ...

Подробнее
01-04-2021 дата публикации

CLIENT-END ENHANCED VIEW PREDICTION FOR MULTI-VIEW VIDEO STREAMING EXPLOITING PRE-FETCHED DATA AND SIDE INFORMATION

Номер: US20210099507A1
Принадлежит:

Techniques for multi-view video streaming are described in the present disclosure, wherein a viewport prediction may be employed at a client-end based on analysis of pre-fetched media item data and ancillary information. A streaming method may first prefetch a portion of content of a multi-view media item. The method may next identify a salient region from the prefetched content and may then download additional content of the media item that corresponds to the identified salient region. 1. A streaming method , comprising:prefetching a portion of content of a multi-view media item,identifying, from the prefetched portion, a salient region of the media item, anddownloading additional content of the media item corresponding to the identified salient region.2. The method of claim 1 , wherein:the prefetched portion is a base layer representation of the multi-view media item,the salient region is identified from decoded base layer data, andthe additional content is an enhancement layer representation of the media item.3. The method of claim 1 , wherein:the prefetched portion is an audio representation of the multi-view media item,the salient region is identified from analysis of the audio representation, andthe additional content is video content of the media item.4. The method of claim 1 , wherein:the prefetched portion is a depth representation of the multi-view media item,the salient region is identified from analysis of the depth representation, andthe additional content is video content of the media item.5. The method of claim 1 , wherein:the prefetched portion is a first point cloud representation of the multi-view media item,the salient region is identified from analysis of the coarse point cloud representation, andthe additional content is a second point cloud representation of the media item at a higher level of detail than the first point cloud representation.6. The method of claim 1 , wherein:the prefetched portion is a first mesh representation of the multi- ...

Подробнее
01-04-2021 дата публикации

Dynamic Point Cloud Compression Using Inter-Prediction

Номер: US20210099711A1
Принадлежит: Apple Inc.

A system comprises an encoder configured to compress attribute information for a dynamic point cloud and/or a decoder configured to decompress compressed attribute information for a dynamic point cloud. The dynamic point cloud may include multiple versions of the point cloud at multiple moments in time Attribute values for the point cloud may be compressed at a reference frame using an intra-prediction process and may be compressed at one or more reference frames using an inter-prediction process that takes advantage of temporal relationships between different frames (e.g. versions) of the dynamic point cloud at the different moments in time. 1. A non-transitory computer-readable medium storing program instructions that , when executed by one or more processors , cause the one or more processors to:compress attribute data for a plurality of points of a point cloud at a first moment in time and at one or more additional moments in time, apply an intra-prediction process at the first moment in time to predict attribute values of the plurality of points at the first moment in time based on predicted or assigned attribute values of neighboring points at the first moment in time;', 'determine residual differences between the predicted attribute values and actual attribute values of the point cloud at the first moment in time; and', 'encode the determined residual differences for the point cloud at the first moment in time,, 'wherein to compress the attribute data at the first moment in time, the program instructions cause the one or more processors to segment the point cloud at the first moment in time into a plurality of segments, each segment comprising one or more points of the point cloud in 3D space;', 'determine motion compensation functions to apply to the segments at the first moment in time to model motion of the points included in the segments at the first moment in time to a target moment in time at one of the one or more additional moments in time;', ' ...

Подробнее
14-04-2016 дата публикации

METADATA HINTS TO SUPPORT BEST EFFORT DECODING FOR GREEN MPEG APPLICATIONS

Номер: US20160105675A1
Принадлежит:

In a coding system, an encoder codes video data according to a predetermined protocol, which, when decoded causes an associated decoder to perform a predetermined sequence of decoding operations. The encoder may perform local decodes of the coded video data, both in the manner dictated by the coding protocol that is at work and also by one or more alternative decoding operations. The encoder may estimate relative performance of the alternative decoding operations as compared to a decoding operation that is mandated by the coding protocol. The encoder may provide identifiers in metadata that is associated with the coded video data to identify such levels of distortion and/or levels of resources conserved. A decoder may refer to such identifiers when determining when to engage alternative decoding operations as may be warranted under resource conservation policies. 1. A method , comprising:coding a video sequence according to a first coding protocol generating a coded video sequence therefrom,decoding the video sequence according to the first coding protocol,decoding the video sequence according to an alternate coding protocol,comparing decoding performance of the first coding protocol to decoding performance of the alternate coding protocol,transmitting, in a channel, coded video data representing the video sequence coded according to the first protocol, and an indicator of relative performance of the alternate coding protocol.2. The method of claim 1 , wherein the comparison of decoding performances includes estimating resource conservation to be achieved by decode of the coded video data according to the alternate coding protocol.3. The method of claim 1 , wherein the comparison of decoding performances includes estimating relative distortion between decode of the coded video data according to the alternate coding protocol and decode of the coded video data according to the first coding protocol.4. The method of claim 1 , further comprisingestimating state of a ...

Подробнее
08-04-2021 дата публикации

Trimming Search Space For Nearest Neighbor Determinations in Point Cloud Compression

Номер: US20210103780A1
Принадлежит: Apple Inc.

A search space for performing nearest neighbor searches for encoding point cloud data may be trimmed. Ranges of a space filling curve may be used to identify search space to exclude or reuse, instead of generating nearest neighbor search results for at least some of the points of a point cloud located within some of the ranges of the space filling curve. Additionally, neighboring voxels may be searched to identify any neighboring points missed during the trimmed search based on the ranges of the space filling curve. 1. One or more non-transitory , computer-readable storage media , storing program instructions that when executed on or across one or more computing devices cause the one or more computing devices to:group points of a point cloud into one or more groups, wherein the points are grouped based on one or more space filling curve value ranges, wherein a point of the point cloud with a space filling curve value within a given one of the one or more space filling curve value ranges is grouped into a same one of the one or more groups with other ones of the points of the point cloud having a space filling curve value within the given space filling curve value range;determine, for respective ones of the one or more groups of grouped points, bounding volumes that bound the points included in the respective group; and determine, for the one or more groups of grouped points, respective distances between a point of the point cloud for which the neighboring point search is being performed and a bounding volume for the one or more groups of grouped points; and', 'exclude from the neighboring point search those points included in respective ones of the one or more groups of grouped points for which the determined distance to the bounding volume for the respective group exceeds a distance threshold., 'perform a neighboring point search, wherein in performing the neighboring point search the program instructions cause the one or more computing devices to2. The one or more ...

Подробнее
03-07-2014 дата публикации

METHODS AND APPARATUSES FOR ADAPTIVELY FILTERING VIDEO SIGNALS

Номер: US20140185693A1
Принадлежит: MAGNUM SEMICONDUCTOR, INC.

Methods and apparatuses for adaptively filtering video content are disclosed herein. An apparatus may include a filter configured to receive a video signal. The filter may further be configured to perform an up-conversion on a plurality of frames of the video signal to provide a plurality of up-converted frames. The filter may further be configured to filter the plurality of up-converted frames to provide a plurality of filtered frames and to perform a down-conversion operation on the plurality of filtered frames to provide a filtered video signal. 1. An apparatus , comprising:a filter configured to receive a video signal and perform an up-conversion operation on a plurality of frames of the video signal to provide a respective plurality of up-converted frames, the filter further configured to filter the plurality of up-converted frames to provide a plurality of filtered frames and perform a down-conversion operation on the plurality of filtered frames to provide a filtered video signal.2. The apparatus of claim 1 , wherein the filter is further configured to perform the up-conversion operation on the plurality of frames based claim 1 , at least in part claim 1 , on a respective amount of motion to which each of the plurality of frames corresponds.3. The apparatus of claim 1 , further comprising:an encoder coupled to the filter and configured to receive the filtered video signal, the encoder further configured to provide an encoded bitstream based, at least in part, on the filtered video signal.4. The apparatus of claim 1 , wherein the filter is further configured to filter the plurality of up-converted frames using motion-compensated temporal filtering processing.5. The apparatus of claim 1 , wherein the filter is configured to perform the up-conversion operation by de-interlacing the plurality of frames.6. The apparatus of claim 1 , wherein the filter is configured to selectively filter components of each of the plurality of up-converted frames based claim 1 , at ...

Подробнее
08-04-2021 дата публикации

Predictive Coding For Point Cloud Compression

Номер: US20210104075A1
Принадлежит: Apple Inc.

An encoder is configured to compress point cloud geometry information using a prediction tree. Ancestor nodes of a node added to the prediction tree may be used to determine the predicted value of the node to be added according to a prediction technique. The prediction tree may be encoded and may be provided for transmission to a decoder that can regenerate the point cloud. 1. One or more non-transitory , computer-readable storage media , storing program instructions that when executed on or across one or more computing devices cause the one or more computing devices to: select individual ones of the plurality of points to include in the plurality of nodes in the prediction tree;', 'determine respective predicted node values for the individual ones of the points determined from respective prediction techniques applied to one or more ancestor nodes of the individual ones of the points, wherein the ancestor nodes are included in the prediction tree;, 'generate a prediction tree comprising a plurality of nodes that correspond to a plurality of points that make up a point cloud captured from one or more sensors, wherein respective ones of the points comprise spatial information for the point, and wherein, in generating the prediction tree, the program instructions cause the one or more computing devices toencode node information from the prediction tree for the plurality of nodes, wherein the encoded node information comprises, for a given node, an indicator of the respective prediction technique applied to determine the respective predicted node value for the given node; andsend or store the encoded node information.2. The one or more non-transitory claim 1 , computer-readable storage media of claim 1 , wherein the predicted node values are predicted spatial values for the individual ones of the selected points.3. The one or more non-transitory claim 1 , computer-readable storage media of claim 1 , wherein the predicted node values are predicted attribute values for ...

Подробнее
08-04-2021 дата публикации

Low-Latency Encoding Using a Bypass Sub-Stream and an Entropy Encoded Sub-Stream

Номер: US20210105022A1
Принадлежит: Apple Inc

A system comprises an encoder configured to entropy encode a bitstream comprising both compressible and non-compressible symbols. The encoder parses the bitstream into a compressible symbol sub-stream and a non-compressible sub-stream. The non-compressible symbol sub-stream bypass an entropy encoding component of the encoder while the compressible symbol sub-stream is entropy encoded. When a quantity of bytes of entropy encoded symbols and bypass symbols is accumulated a chunk of fixed or known size is formed using the accumulated entropy encoded symbol bytes and the bypass bytes without waiting on the full bitstream to be processed by the encoder. In a complementary manner, a decoder reconstructs the bitstream from the packets or chunks.

Подробнее
08-04-2021 дата публикации

Block-Based Predictive Coding For Point Cloud Compression

Номер: US20210105493A1
Принадлежит: Apple Inc

An encoder is configured to compress point cloud information using a blocks of nodes determined from a prediction tree. A prediction tree is generated for a point cloud. Segments of the prediction tree are identified. The segments are divided into blocks that are predicted by predecessor blocks within the segments. The blocks of the prediction tree may then be encoded and may be provided for transmission to a decoder that can regenerate the point cloud from the blocks of the prediction tree.

Подробнее
26-03-2020 дата публикации

EXTENDING SUPPORTED COMPONENTS FOR ENCODING IMAGE DATA

Номер: US20200099942A1
Принадлежит: Apple Inc.

Support for additional components may be specified in a coding scheme for image data. A layer of a coding scheme that specifies color components may also specify additional components. Characteristics of the components may be specified in the same layer or a different layer of the coding scheme. An encoder or decoder may identify the specified components and determine the respective characteristics to perform encoding and decoding of image data. 1. One or more non-transitory , computer-readable storage media , storing program instructions that when executed on or across one or more computing devices cause the one or more computing devices to implement:identifying one or more components of image data to be encoded that are specified in a same layer of a coding scheme that also specifies a number of color components of the coding scheme;determining a respective one or more characteristics for the identified one or more components from the same layer of the coding scheme or a different layer of the coding scheme; andencoding the identified one or more components according to the respective one or more characteristics along with the specified number of color components in the same layer of a bit stream of encoded image data.2. The one or more non-transitory claim 1 , computer-readable storage media of claim 1 , wherein claim 1 , in determining the respective one or more characteristics for the identified one or more components claim 1 , the program instructions cause the one or more computing devices to implement parsing absolute values for the one or more characteristics from the same layer or the different layer.3. The one or more non-transitory claim 1 , computer-readable storage media of claim 1 , wherein claim 1 , in identifying one or more components of image data to be encoded claim 1 , the program instructions cause the one or more computing devices to implement determining a number for the one or more components and the one or more color components specified in ...

Подробнее
02-04-2020 дата публикации

POINT CLOUD COMPRESSION IMAGE PADDING

Номер: US20200104976A1
Принадлежит: Apple Inc.

A system comprises an encoder configured to compress attribute information and/or spatial for a point cloud and/or a decoder configured to decompress compressed attribute and/or spatial information for the point cloud. The encoder is configured to convert a point cloud into an image based representation. The encoder packs patch images into an image frame and fills empty spaces in the image frame with a padding, wherein pixel values for the padding are determined based on neighboring pixels values such that the padding is smoothed in the image frame. Also, the decoder is configured to generate a decompressed point cloud based on an image based representation of a point cloud. 1. A non-transitory computer-readable medium storing program instructions that , when executed on or across one or more processors , cause the one or more processors to:generate, for a point cloud, one or more projected patch images;pack the generated one or more projected patch images into one or more image frames; 'perform a smoothing operation on pixels of the pad such that pixel values for respective ones of the pixels included in the pad are smoothed based on respective pixel values of a set of neighboring pixels in a neighborhood of a given padded pixel being smoothed; and', 'pad spaces in the one or more image frames that are not covered by the one or more projected patch images with a pad, wherein to determine pixel values for the pad the program instructions cause the one or more processors toencode the one or more image frames including the packed patch images and smoothed padded spaces.2. The non-transitory computer-readable medium of claim 1 , wherein program instructions claim 1 , when executed on the one or more processors claim 1 , cause the one or more processors to:for a given image frame, iteratively solve a global cost function to determine the pixel values for the for respective ones of the pixels included in the pad.3. The non-transitory computer-readable medium of claim 2 , ...

Подробнее
02-04-2020 дата публикации

OCCUPANCY MAP BLOCK-TO-PATCH INFORMATION COMPRESSION

Номер: US20200105024A1
Принадлежит: Apple Inc.

A system comprises an encoder configured to compress attribute information and/or spatial information for a point cloud and/or a decoder configured to decompress compressed attribute and/or spatial information for the point cloud. The encoder is configured to convert a point cloud into an image based representation. The encoder packs patch images into an image frame and fills empty spaces in the image frame with a padding. Various compression strategies may be used to encode an occupancy map and related block-to-patch information indicating which portions of the image frame correspond to which packed patches. Packed image frames comprising patches and padding, along with an encoded occupancy map and related block-to-patch information are sent to a decoder. The decoder is configured to generate a decompressed point cloud based on the packed image frames, compressed occupancy map, and related block-to-patch information. 1. A non-transitory computer-readable medium storing program instructions that , when executed by one or more processors , cause the one or more processors to:determine, for a point cloud, a plurality of patches, wherein each patch comprises points of the point cloud projected on a patch plane;pack patch images corresponding to the plurality of patches into corresponding image frames;determine patch information indicating a bounding box size and position for each of the patch images packed into the corresponding image frames;generate one or more occupancy maps comprising information indicating for respective ones of the image frames, whether portions of the respective ones of the image frames are occupied with one of the patch images or correspond to a padded portion that is not occupied with one of the patch images;determine block-to-patch information indicating, for respective blocks of the image frames, a correspondence between the respective block and one or more patch images that have been packed into the image frames such that a portion of the ...

Подробнее
28-04-2016 дата публикации

Adaptive color space transform coding

Номер: US20160119622A1
Автор: Alexandros Tourapis
Принадлежит: Apple Inc

An encoder system may include an analyzer that analyzes a current image area in an input video to select a transform. A selectable residue transformer, controlled by the analyzer, may perform the selectable transform on a residue image generated from the current image area and a predicted current image area, to generate a transformed residue image. An encoder may encode the transformed residue image to generate output data. The analyzer controls the encoder to encode information to identify the selectable transform and to indicate that the selectable transform for the current image area is different from a transform of a previous image area of the input video. A decoder system may include components appropriate for decoding the output data from the encoder system.

Подробнее
17-07-2014 дата публикации

Hybrid Encoding and Decoding Methods for Single and Multiple Layered Video Coding Systems

Номер: US20140198842A1

Encoding and decoding methods for single and multiple layered video coding systems are provided. Specifically, video information provided to a base layer and one or more enhancement layers can be coded using a picture coding mode and/or a residual coding mode. Selection between coding modes can be performed on a region-by-region basis. 2. The method according to claim 1 , wherein claim 1 , in the performing the picture coding mode claim 1 , the second layer reference pictures of picture information are generated from residual type reference pictures from a residual type reference picture buffer associated with the at least one second layer claim 1 , the method further comprising:generating the second layer reference pictures of picture information from the residual type reference pictures by computing a sum between the residual type reference pictures and the at least one second layer rpu-prediction picture.3. The method according to claim 1 , wherein claim 1 , in the performing the picture coding mode claim 1 , the second layer reference pictures of picture information are from a picture type reference picture buffer associated with the at least one second layer.4. The method according to claim 1 , wherein claim 1 , in the performing the residual coding mode claim 1 , the second layer reference pictures of residual information are generated from picture type reference pictures from a picture type reference picture buffer associated with the at least one second layer claim 1 , the method further comprising:generating the second layer reference pictures of residual information from the picture type reference pictures by taking a difference between the picture type reference pictures and the at least one second layer rpu-prediction picture.5. The method according to claim 1 , wherein claim 1 , in the performing the residual coding mode claim 1 , the second layer reference pictures of residual information are generated from picture type reference pictures from a ...

Подробнее