Title of Invention

METHOD AND DECODER FOR COMPOSING A SCENE

Abstract The present invention relates to a method for composing a scene containing a plurality of objects, an object comprising chrominance and luminance components, a chrominance value being associated with a set of at least two luminance values, wherein said method comprises a step of blending a first object with a second object resulting in a blended object, said step comprising the sub-steps of: generating a luminance component of the blended object from the corresponding luminance components of the first and second objects and from a first composition function, and generating a chrominance component of the blended object from the corresponding chrominance components of the first and second object and from a second composition function, the second composition function depending on a set of associated values of the first composition function. Reference: Fig. 3
Full Text

FIELD OF THE INVENTION
The present invention relates to a method for composing a scene containing a plurality of objects, an object comprising chrominance and luminance components, a chrominance value being associated with a set of at least two luminance values, said method comprising a step of blending a first object with a second object resulting in a blended object.
The invention also relates to a decoder, said decoder implementing said method. Such a method may be used in, for example, a MPEG-4 video communication system.
BACKGROUND OF THE INVENTION
A visual scene can be composed of objects of different type such as natural video, two-dimensional, three-dimensional synthetic images, graphics or text.
A video communication system, like for example a mobile telecommunication system, typically comprises a first mobile phone and a second mobile phone, including an encoder and a decoder, respectively. In order to transmit a visual scene from the first mobile phone to the second mobile phone, the different objects are encoded by the first mobile phone and transmitted independently to the second mobile phone. At the decoder side the scene is recomposed by decoding the different objects and then by rendering the decoded objects. The rendering usually comprises a geometric transformation step able to map the object on the scene and a step of blending the mapped objects together, resulting in a final scene.
An object typically comprises chrominance U and V, and luminance Y components forming the color of the object. During the blending step of a current object N with an object M, a transparency value Nalpha can be associated to the object N, said transparency value influencing the chrominance and luminance values Pyuv of a blending object P such that:
Pyuv = [(255 - Nalpha)*Myuv + Nalpha_y*Nyuv]/255
The transparency of a pixel of a blended object is obtained according to the shape of said pixel. The transparency permits to determine which object is behind or in front of another object in the scene. The shape is associated with a pixel of a frame of a scene. It permits to know if a pixel is outside or inside an object. If it is outside, the shape has a null value, otherwise it has a value equal to 1.
For a pixel having coordinates (x,y), the shape value 0 or 1 corresponding to these coordinates is taken and attributed to the luminance of this pixel. In the case where the

composition takes place in the format 4:2:0 (one chrominance value corresponds to four luminance values as illustrated in Fig. 1, the chrominance values being the circles and the luminance values the crosses), the shape used for the chrominance value U or V(x,y) is the shape of the pixel of coordinates (2x-l, 2y-l), i.e. the top right pixel of a square of 2x2 pixels associated with the chrominance value.
One inconvenient of this solution is that the chrominance of the blended object is false. This results in a bad visual quality for objects having an arbitrary shape. This observation is illustrated in the example of an object constituted of the letter "e" and represented by stripes in Fig. 2a to 2c. Fig. 2a is the original object, Fig. 2c is the result of the composition and Fig. 2b explains how the result is obtained. Each square represents a pixel. If one takes the pixel whose coordinates are (4,2), the pixel whose shape will be used for the chrominance component is the pixel (3,1) as shown in Fig. 2b. The value of this shape is 0, that is to say it is outside the letter "e", and the chrominance value will not be taken into account, as shown in Fig. 2c where the pixel value is represented by dots. The result is false because the chrominance value should have been taken into account and the result should have been the same as in Fig. 2a.
SUMMARY OF THE INVENTION
Accordingly, it is an object of the invention to provide a method and a decoder for composing a scene containing a plurality of objects, an object comprising chrominance and luminance components, a chrominance value being associated with a set of at least two luminance values, said method comprising a step of blending a first object with a second object resulting in a blended object, which gives a better visual quality of said blended object.
To this end, there is provided a method wherein the blending step comprising the sub-steps of:
generating a luminance component of the blended object from the corresponding luminance components of the first and second objects and from a first composition function, and
generating a chrominance component of the blended object from the corresponding chrominance components of the first and second object and from a second composition function, the second composition function depending on a set of associated values of the first composition function.
In addition, there is provided a decoder comprising blending means, said blending means comprising:

luminance generation means for generating a luminance component of the blended object from the corresponding luminance components of the first and second objects and from a first composition function, and
chrominance generation means for a generating chrominance component of the blended object from the corresponding chrominance components of the first and second object and from a second composition function, the second composition function depending on a set of associated values of the first composition function.
As we will see in detail further on, using a set of associated values of the first composition function for the second composition function permits to have a better precision for the chrominance component of an object.
According to a first non-limited embodiment, the first composition function depends on a shape component.
In a non-limited variant of this first non- limited embodiment, a chrominance value is associated with 4 luminance values and 4 shape values, the second composition function being an 'OR' function between the 4 associated shape values.
According to a second non-limited embodiment, the first composition function is based on a transparency component.
In a non-limited variant of this second non-limited embodiment a chrominance value is associated with 4 luminance values and 4 transparency values, the second composition function being an average of the 4 transparency values.
BRIEF DESCRIPTION OF THE DRAWINGS
Additional objects, features and advantages of the invention will become apparent upon reading the following detailed description and upon reference to the accompanying drawings in which:
Fig. 1 illustrates a distribution of the chrominance and luminance components in the case of the 4:2:0 format,
Fig. 2a, 2b and 2c illustrate a result of a chrominance computation of a blended object of a scene according to the prior art, and
Fig. 3 is a block diagram of a decoder.
DETAILED DESCRIPTION OF THE INVENTION
The present invention relates to a method for composing a scene, said scene containing a plurality of objects, an object comprising chrominance and luminance

components, a chrominance value being associated with a set of at least two luminance values.
In the following description, the invention is depicted in the case of a video format 4:2:0, i.e. one chrominance U or V value corresponds to four luminance Y values, the chrominance being sub-sampled in the horizontal and vertical directions, but it will be apparent to a person skilled in the art that the invention is applicable to any video format in which the chrominance component is sub-sampled compared to the luminance component.
In the following description, well-known functions or constructions by a person skilled in the art are not described in detail since they would obscure the invention in unnecessary detail.
Fig.3 represents a general overview of a decoder structure. The same decoding scheme is applied when decoding all the objects of a given session.
The decoder comprises a demultiplexer DEMUX (31), a shape decoder SHDEC (32), a motion decoder MDEC (33) and a texture decoder YUVDEC (34) for decoding the luminance and chrominance components. The reconstructed object is obtained by the right combination of the shape, texture and motion information thanks to a motion compensation circuit MC (35), a reconstruction circuit (36) and a picture memory MEM (37), according to a principle known to a person skilled in the art.
After the decoding of the different objects of a given session, an inverse affine transform, such as rotation, a zoom or a reduction for example, is applied on each object of a scene. To this end, a well-known technique called « backward mapping » is used. The inverse affine transform is applied on the coordinates (x,y) of each pixel of said image. The luminance, chrominance and transparency components of a pixel of the current object are then obtained by an interpolation, a bilinear interpolation for example, between four pixels that are surrounding the inverse transformed pixel having coordinates (X,Y).
After the inverse affine transform, the blending step also called composition is performed. The objects are recursively blended in a predetermined order.
In this step, a transparency value is associated to each object of the scene. Practically, it permits to see which object is behind or in front of another object. Each object has its own luminance Y, chrominance U and V, and transparency alpha values. Blending is done sequentially. Usually, the objects are composed two by two beginning by those on the backward of a scene. For example, if a first object N is over-layered over a second object M





It is to be understood that the present invention is not limited to the aforementioned embodiments and variations and modifications may be made without departing from the spirit and scope of the invention as defined in the appended claims. In the respect, the following closing remarks are made.
It is to be understood that the present invention is not limited to the aforementioned video application on mobile phone. For example it can be applied on a television whenever there are some text message and video.
It is to be understood that the method according to the present invention is not limited to the aforementioned implementation.
There are numerous ways of implementing functions of the method according to the invention by means of items of hardware or software, or both, provided that a single item of hardware or software can carries out several functions. It does not exclude that an assembly of items of hardware or software or both carry out a function, thus forming a single function without modifying the method for composing a scene in accordance with the invention.
Said hardware or software items can be implemented in several manners, such as by means of wired electronic circuits or by means of an integrated circuit that is suitable programmed respectively. The integrated circuit can be contained in a computer or in an encoder. In the second case, the decoder comprises luminance generation means for generating a luminance component of the blended object from the corresponding luminance components of the first and second objects and from a first composition function, and chrominance generation means for a generating chrominance component of the blended object from the corresponding chrominance components of the first and second object and from a second composition function, the second composition function depending on a set of values of the first composition function, as described previously, said means being hardware or software items as above stated.
The integrated circuit comprises a set of instructions. Thus, said set of instructions contained, for example, in a computer programming memory or in a decoder memory may cause the computer or the decoder to carry out the different steps of the composing method.
The set of instructions may be loaded into the programming memory by reading a data carrier such as, for example, a disk. A service provider can also make the set of instructions available via a communication network such as, for example, the Internet.
Any reference sign in the following claims should not be construed as limiting the claim. It will be obvious that the use of the verb "to comprise" and its conjugations do not

exclude the presence of any other steps or elements besides those defined in any claim. The article "a" or "an" preceding an element or step does not exclude the presence of a plurality of such elements or steps.







CLAIMS
1 - A method for composing a scene containing a plurality of objects, an object
comprising chrominance and luminance components, a chrominance value being associated with a set of at least two luminance values, wherein said method comprises a step of blending a first object with a second object resulting in a blended object, said step comprising the sub-steps of:
generating a luminance component of the blended object from the corresponding luminance components of the first and second objects and from a first composition function, and
generating a chrominance component of the blended object from the corresponding chrominance components of the first and second object and from a second composition function, the second composition function depending on a set of associated values of the first composition function.
2. A method for composing a scene as claimed in claim 1, wherein the first composition function is based on a transparency component.
3. A method for composing a scene as claimed in claim 2, wherein a chrominance value is associated with 4 luminance values and 4 transparency values, the second composition function being an average of the 4 transparency values.
4. A method for composing a scene as claimed in claim 1, wherein the first composition function depends on a shape component.
5. A method for composing a scene as claimed in claim 4, wherein a chrominance value is associated with 4 luminance values and 4 shape values, the second composition function being an 'OR' function between the 4 associated shape values.
6. A decoder for composing a scene containing a plurality of objects, an object comprising chrominance and luminance components, a chrominance value being associated with a set of at least two luminance values, said decoder comprising means for blending a first object with a second object resulting in a blended object, said blending means comprising:

luminance generation means for generating a luminance component of the blended object from the corresponding luminance components of the first and second objects and from a first composition function, and
chrominance generation means for a generating chrominance component of the blended object from the corresponding chrominance components of the first and second object and from a second composition function, the second composition function depending on a set of associated values of the first composition function.
7. A computer program product comprising program instructions for
implementing, when said program is executed by a processor, a composition method as claimed in claim 1.


Documents:

3497-chenp-2005 abstract duplicate.pdf

3497-chenp-2005 claims duplicate.pdf

3497-chenp-2005 description(complete) duplicate.pdf

3497-chenp-2005 drawings duplicate.pdf

3497-CHENP-2005 FORM 13.pdf

3497-CHENP-2005 FORM 18.pdf

3497-CHENP-2005 FORM 6.pdf

3497-chenp-2005-abstract.pdf

3497-chenp-2005-claims.pdf

3497-chenp-2005-correspondnece-others.pdf

3497-chenp-2005-description(complete).pdf

3497-chenp-2005-drawings.pdf

3497-chenp-2005-form 1.pdf

3497-chenp-2005-form 3.pdf

3497-chenp-2005-form 5.pdf

3497-chenp-2005-pct.pdf


Patent Number 231301
Indian Patent Application Number 3497/CHENP/2005
PG Journal Number 13/2009
Publication Date 27-Mar-2009
Grant Date 04-Mar-2009
Date of Filing 22-Dec-2005
Name of Patentee NXP B.V
Applicant Address HIGH TECH CAMPUS 60, 5656 AG EINDHOVEN,
Inventors:
# Inventor's Name Inventor's Address
1 GOBERT, Jean c/o Société Civile SPID, 156 Boulevard Haussmann, F-75008 PARIS,
2 NGUYEN-PHUC, Lien Société Civile SPID, 156 Boulevard Haussmann, F-75008 PARIS,
PCT International Classification Number H04N7/26
PCT International Application Number PCT/IB04/01999
PCT International Filing date 2004-06-08
PCT Conventions:
# PCT Application Number Date of Convention Priority Country
1 03300036.5 2003-06-23 EUROPEAN UNION