白丝美女被狂躁免费视频网站,500av导航大全精品,yw.193.cnc爆乳尤物未满,97se亚洲综合色区,аⅴ天堂中文在线网官网

Apparatus, a method and a computer program for video coding and decoding

專利號(hào)
US11800131B2
公開日期
2023-10-24
申請(qǐng)人
NOKIA TECHNOLOGIES OY(FI Espoo)
發(fā)明人
Miska Matias Hannuksela; Kemal Ugur
IPC分類
H04N19/463; H04N19/70; H04N19/30; H04N19/61
技術(shù)領(lǐng)域
prediction,picture,bitstream,layer,pictures,syntax,coded,in,decoding,nal
地域: Espoo

摘要

A method comprising encoding a bitstream comprising a base layer, a first enhancement layer and a second enhancement layer; encoding an indication of both the base layer and the first enhancement layer used for prediction for the second enhancement layer in the bitstream; encoding, in the bitstream, an indication of a first set of prediction types that is applicable from the base layer to the second enhancement layer, wherein the first set of prediction types is a subset of all prediction types available for prediction between layers, and encoding, in the bitstream, an indication of a second set of prediction types that is applicable from the base layer or the first enhancement layer to the second enhancement layer, wherein the second set of prediction types is a subset of all prediction types available for prediction between layers.

說明書

CROSS-REFERENCE TO RELATED APPLICATIONS

This application is a continuation of U.S. application Ser. No. 16/689,582, filed Nov. 20, 2019, which is a continuation of U.S. application Ser. No. 15/899,129, filed Feb. 19, 2018, which is a continuation of U.S. application Ser. No. 14/143,986, filed Dec. 30, 2013, which claims priority to U.S. Provisional Application No. 61/748,938, filed Jan. 4, 2013, the entire contents of which are incorporated herein by reference.

TECHNICAL FIELD

The present invention relates to an apparatus, a method and a computer program for video coding and decoding.

BACKGROUND

A video codec may comprise an encoder which transforms input video into a compressed representation suitable for storage and/or transmission and a decoder that can uncompress the compressed video representation back into a viewable form, or either one of them. Typically, the encoder discards some information in the original video sequence in order to represent the video in a more compact form, for example at a lower bit rate.

Scalable video coding refers to coding structure where one bitstream can contain multiple representations of the content at different bitrates, resolutions or frame rates. A scalable bitstream typically consists of a “base layer” providing the lowest quality video available and one or more enhancement layers that enhance the video quality when received and decoded together with the lower layers. In order to improve coding efficiency for the enhancement layers, the coded representation of that layer typically depends on the lower layers.

權(quán)利要求

1
That which is claimed is:1. A method comprising:encoding a bitstream comprising a base layer, a first enhancement layer and a second enhancement layer;encoding, in the bitstream, an indication of a number of bits in a prediction type mask syntax element;encoding, in the bitstream using a first prediction type mask syntax, an indication of a first set of prediction types that is applicable from the base layer to the second enhancement layer, wherein the first set of prediction types is a subset of all prediction types available for prediction between layers; andencoding, in the bitstream using a second prediction type mask syntax, an indication of a second set of prediction types that is applicable from the first enhancement layer to the second enhancement layer, wherein the second set of prediction types is a subset of all prediction types available for prediction between layers,wherein each of said prediction types available for prediction between layers is represented in the first prediction type mask syntax and the second prediction type mask syntax, andwherein said prediction types available for prediction between layers are adaptively selectable as at least one of the following: sample prediction, motion information prediction or filtering parameter prediction.2. The method according to claim 1, further comprising:encoding a picture of the base layer and a picture of the first enhancement layer; andencoding a picture of said second enhancement layer using said first set of prediction types from the picture of the base layer and said second set of prediction types from the picture of the first enhancement layer.3. The method according to claim 1,wherein each of said prediction types available for prediction between layers is represented by a bit number in the first prediction type mask syntax and the second prediction type mask syntax.4. The method according to claim 1, wherein said indication of the first set of prediction types and said indication of the second set of prediction types are included in at least one of a sequence parameter set or a video parameter set.5. The method according to claim 1, further comprising encoding, in the bitstream, an indication of at least one set of prediction types that is not applicable from the base layer or the first enhancement layer to the second enhancement layer.6. The method according to claim 1, wherein the second enhancement layer enhances a first scalability type relative to the base layer and a second scalability type relative to the first enhancement layer, and wherein the first scalability type and second scalability type are selected from at least one of: temporal scalability, quality scalability, spatial scalability, view scalability, depth enhancements, bit-depth scalability, chroma format scalability or color gamut scalability.7. The method according to claim 6, wherein the prediction types available for prediction between the second enhancement layer and the base layer are dependent on the first scalability type, and wherein the prediction types available for prediction between the second enhancement layer and the first enhancement layer are dependent on the second scalability type.8. The method according to claim 1, wherein the first set of prediction types has a first prediction direction and the second set of prediction types has a second prediction direction, and wherein said first prediction direction and second prediction direction are one of the following: temporal prediction, inter-view prediction, inter-layer prediction or inter-component prediction.9. An apparatus comprising:at least one processor and at least one memory, said at least one memory stored with code thereon, which when executed by said at least one processor, causes the apparatus to perform:encoding a bitstream comprising a base layer, a first enhancement layer and a second enhancement layer;encoding, in the bitstream, an indication of a number of bits in a prediction type mask syntax element;encoding, in the bitstream using a first prediction type mask syntax, an indication of a first set of prediction types that is applicable from the base layer to the second enhancement layer, wherein the first set of prediction types is a subset of all prediction types available for prediction between layers; andencoding, in the bitstream using a second prediction type mask syntax, an indication of a second set of prediction types that is applicable from the first enhancement layer to the second enhancement layer, wherein the second set of prediction types is a subset of all prediction types available for prediction between layers,wherein each of said prediction types available for prediction between layers is represented in the first prediction type mask syntax and the second prediction type mask syntax, andwherein said prediction types available for prediction between layers are adaptively selectable as at least one of the following: sample prediction, motion information prediction or filtering parameter prediction.10. The apparatus according to claim 9, wherein the apparatus is also caused to:encode a picture of the base layer and a picture of the first enhancement layer; andencode a picture of said second enhancement layer using said first set of prediction types from the picture of the base layer and said second set of prediction types from the picture of the first enhancement layer.11. The apparatus according to claim 9,wherein each of said prediction types available for prediction between layers is represented by a bit number in the first prediction type mask syntax and the second prediction type mask syntax.12. The apparatus according to claim 9, wherein said indication of the first set of prediction types and said indication of the second set of prediction types are included in at least one of a sequence parameter set or a video parameter set.13. The apparatus according to claim 9, wherein the apparatus is further configured to encode, in the bitstream, an indication of at least one set of prediction types that is not applicable from the base layer or the first enhancement layer to the second enhancement layer.14. The apparatus according to claim 9, wherein the second enhancement layer enhances a first scalability type relative to the base layer and a second scalability type relative to the first enhancement layer, and wherein the first scalability type and second scalability type are selected from at least one of: temporal scalability, quality scalability, spatial scalability, view scalability, depth enhancements, bit-depth scalability, chroma format scalability or color gamut scalability.15. The apparatus according to claim 14 wherein the prediction types available for prediction between the second enhancement layer and the base layer are dependent on the first scalability type, and wherein the prediction types available for prediction between the second enhancement layer and the first enhancement layer are dependent on the second scalability type.16. The apparatus according to claim 9, wherein the first set of prediction types has a first prediction direction and the second set of prediction types has a second prediction direction, and wherein said first prediction direction and second prediction direction are one of the following: temporal prediction, inter-view prediction, inter-layer prediction or inter-component prediction.17. A non-transitory computer readable storage medium stored with code thereon for use by an apparatus, which when executed by a processor, causes the apparatus to perform:encoding a bitstream comprising a base layer, a first enhancement layer and a second enhancement layer;encoding, in the bitstream, an indication of a number of bits in a prediction type mask syntax element;encoding, in the bitstream using a first prediction type mask syntax, an indication of a first set of prediction types that is applicable from the base layer to the second enhancement layer, wherein the first set of prediction types is a subset of all prediction types available for prediction between layers; andencoding, in the bitstream using a second prediction type mask syntax, an indication of a second set of prediction types that is applicable from the first enhancement layer to the second enhancement layer, wherein the second set of prediction types is a subset of all prediction types available for prediction between layers,wherein each of said prediction types available for prediction between layers is represented in the first prediction type mask syntax and the second prediction type mask syntax, andwherein said prediction types available for prediction between layers are adaptively selectable as at least one of the following: sample prediction, motion information prediction or filtering parameter prediction.18. The computer readable storage medium according to claim 17, wherein the code, when executed by the processor, also causes the apparatus to:encode a picture of the base layer and a picture of the first enhancement layer; andencode a picture of said second enhancement layer using said first set of prediction types from the picture of the base layer and said second set of prediction types from the picture of the first enhancement layer.19. The computer readable storage medium according to claim 17,wherein each of said prediction types available for prediction between layers is represented by a bit number in the first prediction type mask syntax and the second prediction type mask syntax.20. The computer readable storage medium according to claim 17, wherein said indication of the first set of prediction types and said indication of the second set of prediction types are included in at least one of a sequence parameter set or a video parameter set.21. The computer readable storage medium according to claim 17, wherein the code, when executed by the processor, also causes the apparatus to encode, in the bitstream, an indication of at least one set of prediction types that is not applicable from the base layer or the first enhancement layer to the second enhancement layer.22. The computer readable storage medium according to claim 17, wherein the second enhancement layer enhances a first scalability type relative to the base layer and a second scalability type relative to the first enhancement layer, and wherein the first scalability type and second scalability type are selected from at least one of: temporal scalability, quality scalability, spatial scalability, view scalability, depth enhancements, bit-depth scalability, chroma format scalability or color gamut scalability.23. The computer readable storage medium according to claim 22, wherein the prediction types available for prediction between the second enhancement layer and the base layer are dependent on the first scalability type, and wherein the prediction types available for prediction between the second enhancement layer and the first enhancement layer are dependent on the second scalability type.24. A method comprising:decoding, from a bitstream, an indication of a number of bits in a prediction type mask syntax element;decoding, from a first prediction type mask syntax in the bitstream, an indication of a first set of prediction types that is applicable from a base layer to a second enhancement layer, wherein the first set of prediction types is a subset of all prediction types available for prediction between layers;decoding, from a second prediction type mask syntax in the bitstream, an indication of a second set of prediction types that is applicable from a first enhancement layer to the second enhancement layer, wherein the second set of prediction types is a subset of all prediction types available for prediction between layers;decoding a picture of the base layer and a picture of the first enhancement layer; anddecoding a picture of said second enhancement layer using said first set of prediction types from the picture of the base layer and said second set of prediction types from the picture of the first enhancement layer,wherein each of said prediction types available for prediction between layers is represented in the first prediction type mask syntax and the second prediction type mask syntax, andwherein said prediction types available for prediction between layers are at least one of the following: sample prediction, motion information prediction or filtering parameter prediction.25. The method according to claim 24,wherein each of said prediction types available for prediction between layers is represented by a bit number in the first prediction type mask syntax and the second prediction type mask syntax.26. The method according to claim 24, wherein said indication of the first set of prediction types and said indication of the second set of prediction types are decoded from at least one of a sequence parameter set or a video parameter set.27. The method according to claim 24, further comprising decoding, from the bitstream, an indication of at least one set of prediction types that is not applicable form the base layer or the first enhancement layer to the second enhancement layer.28. The method according to claim 24, wherein the second enhancement layer enhances a first scalability type relative to the base layer and a second scalability type relative to the first enhancement layer, and wherein the first scalability type and second scalability type are selected from at least one of: temporal scalability, quality scalability, spatial scalability, view scalability, depth enhancements, bit-depth scalability, chroma format scalability or color gamut scalability.29. The method according to claim 28, wherein the prediction types available for prediction between the second enhancement layer and the base layer are dependent on the first scalability type, and wherein the prediction types available for prediction between the second enhancement layer and the first enhancement layer are dependent on the second scalability type.30. The method according to claim 24, wherein the first set of prediction types has a first prediction direction and the second set of prediction types has a second prediction direction, and wherein said first prediction direction and second prediction direction are one of the following: temporal prediction, inter-view prediction, inter-layer prediction or inter-component prediction.31. An apparatus comprising:at least one processor and at least one memory, said at least one memory stored with code thereon, which when executed by said at least one processor, causes the apparatus to perform:decoding, from a bitstream, an indication of a number of bits in a prediction type mask syntax element;decoding, from a first prediction type mask syntax in the bitstream, an indication of a first set of prediction types that is applicable from a base layer to a second enhancement layer, wherein the first set of prediction types is a subset of all prediction types available for prediction between layers;decoding, from a second prediction type mask syntax in the bitstream, an indication of a second set of prediction types that is applicable from a first enhancement layer to the second enhancement layer, wherein the second set of prediction types is a subset of all prediction types available for prediction between layers;decoding a picture of the base layer and a picture of the first enhancement layer; anddecoding a picture of said second enhancement layer using said first set of prediction types from the picture of the base layer and said second set of prediction types from the picture of the first enhancement layer,wherein each of said prediction types available for prediction between layers is represented in the first prediction type mask syntax and the second prediction type mask syntax, andwherein said prediction types available for prediction between layers are at least one of the following: sample prediction, motion information prediction or filtering parameter prediction.32. The apparatus according to claim 31,wherein each of said prediction types available for prediction between layers is represented by a bit number in the first prediction type mask syntax and the second prediction type mask syntax.33. The apparatus according to claim 31, wherein said indication of the first set of prediction types and said indication of the second set of prediction types are decoded from at least one of a sequence parameter set or a video parameter set.34. The apparatus according to claim 31, wherein the apparatus is further caused to decode, from the bitstream, an indication of at least one set of prediction types that is not applicable form the base layer or the first enhancement layer to the second enhancement layer.35. The apparatus according to claim 31, wherein the second enhancement layer enhances a first scalability type relative to the base layer and a second scalability type relative to the first enhancement layer, and wherein the first scalability type and second scalability type are selected from at least one of: temporal scalability, quality scalability, spatial scalability, view scalability, depth enhancements, bit-depth scalability, chroma format scalability or color gamut scalability.36. The apparatus according to claim 35, wherein the prediction types available for prediction between the second enhancement layer and the base layer are dependent on the first scalability type, and wherein the prediction types available for prediction between the second enhancement layer and the first enhancement layer are dependent on the second scalability type.37. The apparatus according to claim 31, wherein the first set of prediction types has a first prediction direction and the second set of prediction types has a second prediction direction, and wherein said first prediction direction and second prediction direction are one of the following: temporal prediction, inter-view prediction, inter-layer prediction or inter-component prediction.38. A non-transitory computer readable storage medium stored with code thereon for use by an apparatus, which when executed by a processor, causes the apparatus to perform:decoding, from a bitstream, an indication of a number of bits in a prediction type mask syntax element;decoding, from a first prediction type mask syntax in the bitstream, an indication of a first set of prediction types that is applicable from a base layer to a second enhancement layer, wherein the first set of prediction types is a subset of all prediction types available for prediction between layers;decoding, from a second prediction type mask syntax in the bitstream, an indication of a second set of prediction types that is applicable from a first enhancement layer to the second enhancement layer, wherein the second set of prediction types is a subset of all prediction types available for prediction between layers;decoding a picture of the base layer and a picture of the first enhancement layer; anddecoding a picture of said second enhancement layer using said first set of prediction types from the picture of the base layer and said second set of prediction types from the picture of the first enhancement layer,wherein each of said prediction types available for prediction between layers is represented in the first prediction type mask syntax and the second prediction type mask syntax, andwherein said prediction types available for prediction between layers are at least one of the following: sample prediction, motion information prediction or filtering parameter prediction.39. The computer readable storage medium according to claim 38,wherein each of said prediction types available for prediction between layers is represented by a bit number in the first prediction type mask syntax and the second prediction type mask syntax.40. The computer readable storage medium according to claim 38, wherein said indication of the first set of prediction types and said indication of the second set of prediction types are decoded from at least one of a sequence parameter set or a video parameter set.41. The computer readable storage medium according to claim 38, wherein the code, when executed by the processor, also causes the apparatus to decode, from the bitstream, an indication of at least one set of prediction types that is not applicable form the base layer or the first enhancement layer to the second enhancement layer.42. The computer readable storage medium according to claim 38, wherein the second enhancement layer enhances a first scalability type relative to the base layer and a second scalability type relative to the first enhancement layer, and wherein the first scalability type and second scalability type are selected from at least one of: temporal scalability, quality scalability, spatial scalability, view scalability, depth enhancements, bit-depth scalability, chroma format scalability or color gamut scalability.43. The computer readable storage medium according to claim 42, wherein the prediction types available for prediction between the second enhancement layer and the base layer are dependent on the first scalability type, and wherein the prediction types available for prediction between the second enhancement layer and the first enhancement layer are dependent on the second scalability type.
微信群二維碼
意見反饋