Title of Invention

A METHOD FOR TRANSMITTING DATA FROM A PLURALITY OF DATA PROCESSING DEVICES ACROSS A DATA PACKET COMMUNICATIONS NETWORK

Abstract The invention provides a method for transmitting data from a plurality of data sources across data packet data communications network having a congestion control mechanism for reducing the effects of congestion by selective 1> prioritising data packets. The data packets can contain data in a number oi different multimedia types, e.g. voice, video, audio, email each being within a separate partition in the packet. The packets can be transmitted as a data packet train, which consists of a number of data packets with some association in time and an order of precedence. The association and order of precedence are used to decide which packets can be kept and which packets can be discarded in the presence of a congested network. The data packet partitioning may be mack-adaptive where the lengths of data packet partitions can be varied dynamical 1> according to the type of data present and current network conditions.
Full Text

Data packet Node, and Methofl of Operating a Data Packet Network
Field of the Invention
This invention relates to data packet nodes, and methods of operating a data packet network, incorporating quality control mechanisms for the transmission of data across the network, and in particular for the transmission of data across a network having a congestion control mechanism for reducing the effect of network congestion by selectively prioritising data packets.
Background of the Invention
A problem with conventional data packet networks is that their operation is based upon a 'best effort* paradigm: a data packet is presented to the network without the certainty that it will be delivered. There are no a-priori agreements between the sender and receiver of the data packet to ensure such certainty. However, various techniques have been developed to support quality management of data packet networks, typically including dedicated bandwidth allocation and/or congestion control mechanisms for reducing the effect of network congestion by selectively prioritising data packets. Such congestion control mechanisms include systems where certain data packets can be tagged, to give them priority in their handling over other data packets, or in their tendency not to be discarded, relative to others within the system of lower precedence.
United States patent 5,541,919, describes data source segmentation and multiplexing in a multimedia communications system. Packet segmentation and multiplexing are performed dynamically based on the fullness of a set of information buffers and the delay sensitivity of each data source.
An international workshop submission by Toufik Ahmed et ah, of the University of Versailles, entitled "Adaptive MPEG-4 Streaming based on AVO Classification and Network Congestion Feedback" relates to an adaptive object orientated streaming framework for a unicast Moving Picture Expert Group version 4 (MPEG-4) stream over a Transmission Control Protocol - Internet Protocol (TCP-IP) network, with particular application to a Video on Demand

(VoD) service. In the MPEG-4 standard, video scenes are encoded using object-based compression where different audio visual objects (AVOs) in a scene aie encoded separately. The submission employs a streaming server to classify the AVOs using certain application-level QoS criteria and also according to their importance to a scene. The more itnportant AVOs are streamed before the less important ones and the streaming server deals with network congestion by halting the streaming of less important AVOs when congestion is detected.
European patent application EP 0 544 452 describes a system in which core information, for example in the form of a core block or blocks, is transmitted in a core packet, and at least some enhancement information, for example, in the form of enhancement blocks, is transmitted in an enhancement packet which is separate from the core packet and is discardable to relieve congestion. The core and enhancement packets may have headers which include a discard eligible marker to indicate whether or not the associated packet can be discarded. The enhancement blocks may be distributed between the core packet and enhancement packet in accordance with congestion conditions, or the enhancement blocks may be incorporated only in the enhancement packet, and the actual number of enhancement blocks included are varied depending on congestion conditions. The system preserves at least some form of service for a voice signal by dropping enhancement layer data packets from the voice signal during periods of congestion in the network.
A method of operating a data packet network to provide selectable levels of service to different communication flows is disclosed in International patent application WO 02/071702.
Two important works tackling real-time Quality of Service (QoS) in a data packet network are the IntServ and DiffServ approaches, described in R. Braden, et alt "Integrated Services in the Internet Architecture: an Overview," KFC1633, Jun 1994 and K. Nichols, et al, "Definition of the Differentiated Services field in the IPv4 and IPv6 headers," RFC, Dec. 1998, respectively. The former architecture satisfied both necessary conditions for the network QoS i.e. it provided appropriate bandwidth and queuing resources for each application flow. However, the additional complexity involved in the implementation of the

hop signalling renders the process unscalable for public network operation. The latter architecture incorporates queue servicing mechanisms with scheduling and data packet discarding, but does not guarantee bandwidth and thus satisfies only the second necessary condition for QoS.
In United States patent application US 2002/0181506, a scheme for supporting real-time data packetisation of multimedia information is disclosed. The scheme involves storing copies of transmission data packets for a predetermined time period and rescinding upon detection of lost data packets. The scheme further involves reading a stream into memory prior to processing and therefore cannot be described as true real-time.
A problem common to data packet networks which have congestion control mechanisms which prioritise some data packets over others is that, whilst they enable high priority traffic to be delivered, this is at the expense of low priority traffic. At times of high congestion, this can result in no low priority traffic arriving at the destination.
Another common problem in data packet networks are the delays incurred through the network. Certain data sources have strict time intervals in which their data must arrive at their destination. In order to increase tolerance to delay, it would be desirable to have the facility to prepare resources in advance of data reception.
Summary of the Invention
In accordance with a first aspect of the present invention, there is provided a method for transmitting data from a plurality of data processing devices across a data packet data communications network having a congestion control mechanism for reducing the effects of congestion by selectively prioritising data packets, the method comprising the steps of:
receiving data from at least a first data processing device and a second data processing device;
constructing a first data packet for carrying data through said network;
constructing a second data packet for carrying data through said network; attaching prioritisation infoimation to at least one of the first and

second data packets, the prioritisation information being for use by the congestion control mechanism to prioritise the fust data packet in preference to the second data packet; and
transmitting the first and second data packets into said network, characterised in that the first packet construction process comprises adding data from both the first data processing device and the second data processing device to the first data packet in controlled amounts, the amount of data from each of the first and second data processing devices added to the first packet being controlled during the first packet construction process; and
the second packet construction process comprises adding data from at least one of the first and second data processing devices to the second data packet.
Hence, by use of the present invention, even if a second data packet containing data from one or more data processing devices is discarded on its route through the network, it is still possible to deliver an acceptable level of service for two or more data processing devices by delivery of a first data packet containing data from two or more data processing devices. This scheme can clearly be extended to a higher number of data processing devices and data packets, providing further levels of service.
In accordance with a second aspect of the present invention, there is provided a method of transmitting data using a plurality of different data formats across a data packet data communications network, the method comprising the steps of:
selecting a first data format from said plurality of data formats;
adding data to a first data packet, in the first data format;
transmitting the first data packet into the network;
selecting a second, different format from the plurality of data formats;
adding data to a second data packet, in the second data format; and
transmitting the second data packet into the network,
characterised in that before said first data packet is transmitted into the network, advance warning data of the format of said second data packet to be

constructed subsequently to said first data packet is added into the first data packet.
By use of the present invention, it is possible to alter the contents of data packets according to present traffic levels and also incorporate advance warning data into the data packets. The advance warning data contains information on data packets to be sent subsequently and can be used by the destination to prepare in advance for the reception of data packets. Such advance warning will inherently enable resources to be more efficiently used and hence reduce delay through the system.
In accordance with a third aspect of the present invention, there is provided a method for transmitting data from a plurality of data processing devices across a data packet data communications network, the method comprising the steps of:
receiving data from at least a first data processing device and a second data processing device;
constructing data packets for carrying data through said network, characterised in that the packet construction process comprises adding data from both the first data processing device and the second data processing device to the first data packet in controlled amounts, the amount of data from each of the first and second data processing devices added to the first packet being controlled during the first packet construction process; and
the relative proportions of data from the first and second data processing devices in the data packets are varied in dependence on current conditions of transmission of data through the network-In preferred embodiments, this aspect of the invention provides for the dynamic partitioning of packets based on current network conditions.
Further features and advantages of the invention will become apparent from the following description of preferred embodiments of the invention, given by way of example only, which is made with reference to the accompanying drawings.

Brief Description of the Drawings
Figure 1 is an overall system diagram of an example data packet switched communication network.
Figure 2 is a schematic illustration of a data packet train transmitter according to an embodiment of the invention.

Figure 3 is a schematic illustration of the partitioning of three data packet payloads of a data packet train according to an embodiment of the invention.
Detailed Description of the Inventjoi^
An overall system diagram according to an embodiment of the invention is shown in Figure 1. This gives an example of a communications system where the present invention could be applied, but is by no means the only scenario of application. A set of data processing devices, 9ยป 10, 11, are shown on the left hand side of the diagram. These devices could include one or more of a wireless device 9, such as a cellular telephone, personal digital assistant (PDA), laptop computer, etc., a computer workstation 10 and/or a server computer 11. The devices produce different types of daHa, SI, S2> S3, which are received by a first network edge node 12 e.g. a cellular comrotinicatians network "base station.
The data is passed on through a first data packet communications network 14 such as a mobile communications data packet network, for example a General Data packet Radio Network (GPRS). The data is then communicated via a second data packet communications network 16, for example an interact backbone network, to a second network edge node 18. The data is then passed from the second edge node 18 on to at least one of a variety of data processing devices 20, 22, 24 similar to the wireless device 9, computer workstation 10 or server computer 11 mentioned above.
The present invention provides improved data transmission mechanisms, which may be implemented in the first network edge node 12, whereby information caa be transmitted through the data packet network infrastructure elements 14, 16 and received at the second network edge node 18. This is indicated on Figure 1 by the dotted arrow 26.
The invention provides three new and interrelated features which may be implemented in the first network edge node to support synchronised multimedia data packet traffic: 1. The transmission of data using mixed multi-media ("MMM") data packet
trains

2. The transmission of MMM data packets having a priori knowledge of the
format of subsequent data packets; and
3. Adaptive MMM data packet partitioning.
MMM Data Packet Trains
An MMM data packet is a data packet that can contain data in a mixture of multimedia types. These multimedia types could be voice, video, audio, email, etc. Some types of multimedia data can have the requirement of real-time operation, in applications such as voice calls, video conferencing and radio. The other types, such as email, are not intended for real-time use and are referred to herein as asynchronous data types. There is then, a need to distinguish between these different data types and handle their routeing accordingly.
In the preferred embodiment of the present invention, transcoders are employed to convert data into a format suitable for being sent across a data packet network based upon the congestion characteristic at that point in time. The data is then data packetised into data packet trains, each data packet train including,a plurality of data packets and each of the plurality of data packets including data from at least one of the sources. The data packets within a train need not necessarily be sent together, travel through the network together or arrive together.
A data packet fain is defined as a set of data packets that have an association in time, and an order of precedence. MMM data packet trains are formed sequentially, such that respective data packet trains are created using source data received, and transmitted, during a respective and sequential periods of time. There must be a minimum of two data packets in a train to form an association between them, but the upper limit is undefined and would be determined by the particular implementation and type of data passing through it. A physical constraint on the size of a data packet train is the total amount of information that can be stored in the buffers.
A data packet train transmitter system according to one embodiment of the present invention is shown in Figure 2. A number of input data sources 100, 101. etc. are fed into a number of transcoders 102A. 102B. 102O: 103A. 103R

etc. In Figure 2, only two input data sources, SI and S2, are shown, but it should be appreciated that more are possible in practice. Similarly, only a given number of transcoders are shown, but there also can be many more. The transcoders then feed the data on to a plurality of buffers 105, 106, 107, of which there is at least one for each source SI, S2, etc., which hold the data until requested by the data packet partition loader 108.
The buffer monitor 122 provides information to the transcoder selector 118 in response to detecting a predetermined fill level of the buffers, to indicate which buffers are becoming full. The transcoder selector 118 uses this information to select which of the transcoders 102, 104 to use for the data to be transcoded next. The transcoder selector 118 also feeds information about a change of transcoder affecting a subsequent data packet on to the payload header constructor 110 via an advance warning loader 120 so that this information can be added to the data packet header to reduce system delay in the reverse transcoding process at the second network edge node 18. Once the data packet partition loader 108 has loaded the data packet partitions, the payload header constructor 110 adds a MMM data packet header to each data packet.
Control of the data packet partition loader 108 and the payload header constructor 110 is carried out by a dynamic payload controller 114 which decides on the partition length and contents of each data packet. The number and order of data packets in a train is then calculated by the data packet train sequencer 116 which informs the payload header constructor 110 of its decisions, so that this information can also be added to the MMM data packet headers. Finally, a packetiser 112 is used to create the completed data packets by appending a transport protocol header to form each MMM data packet, so that they can be transmitted into the existing network infrastructure with suitable routeing information indicating the destination of the data, which in this embodiment is the second network edge node 18. At the second network edge node, the data from each of the sources in the MMM data packet train is separately reconstructed and forwarded to the suitable receiving terminal 20, 22 or 24.

At least one of, and preferably all of, the data packets in an MMM data packet train are divided into several partitions of different length, as shown in Figure 3, with boundaries 40 between the partitions containing data from each different data source. In the embodiment shown in Figure 3, the MMM data packet train includes a first data packet 42, a second data packet 44 and a third data packet 46.
The contents of each partition in each data packet are taken from different respective data sources SI, S2 and S3. The packet partition loader 108 allocates each source an associated level of importance; in the embodiment shown, data source SI has the highest level of importance, followed by S2, and S3 has the lowest level of importance. The packet partition loader 108 uses this relative importance hierarchy to determine the amounts of data from each source to be included in each different packet in the MMM data packet train. In the first packet 42, the packet partition loader 108 includes a relatively high proportion of data from the first source SI, a lesser proportion of data from the second source S2, and relatively low proportion of data from the third source S3. In the second packet 44, the packet partition loader 108 includes, relative to the amounts included in the first packet 42, a lower proportion of data from the first source SI, a higher proportion of data from the second source S29 and a higher proportion of data from the third source S3. In the third packet 46, the packet partition loader 108 includes, relative to the amounts included in the second packet 44? a lower proportion of data from the first source SI, a higher proportion of data from the second source S2, and a higher proportion of data from the third source S3. Moreover, in the third packet 46, the packet partition loader 108 includes a relatively low proportion of data from the first source SI, a higher proportion of data from the second source S2, and relatively high proportion of data from the third source S3.
Note that regions 72, 78 and 84 together constitute data from SI. Similarly regions 74, 80 and 86 together constitute data from S2 and regions 76, 82 and 88 together constitute data from S3. Note that the amount of data from each source included in a packet train is preferably less than then buffer size of the respective source buffer 105, 106, 107, so that the maximum amount of data

from each source in the packet train is constrained by the maximum contents of the respective source buffer 105,106, 107.
The different data types may each be given an importance value in dependence on their tolerance to delay, where a least delay-tolerant data type is given the highest priority and a most delay-tolerant data type is given the lowest priority. If two or more data types have an equal delay tolerance, they may be given the same priority level and be grouped into a single priority group. The importance level may also, or alternatively, be based on other factors, such as the importance value of the content of the data type e.g. one data source may be carrying data that has to be delivered for some form of emergency or data which deemed to have no tolerance to delivery failure, such as financial transaction information.
In a preferred embodiment of the invention, each MMM data packet will also contain a MMM header part in the payload, containing information about what data the data packet contains and how the data packet has been partitioned. This header may be located anywhere within the data packet payload, although as shown in the preferred embodiment of Figure 3, the payload 48 consists of data from the various sources SI, S2, S3 and the MMM data packet header at its head.
A farther header in the form of a transport protocol header 60, 64, 68, is then added at the front of the MMM data packet. This transport protocol header could be the form of known Internet Protocol (IP) or X.25 protocol headers. Typically, the transport protocol header contains such information as source and destination address, time stamp, length and type of service etc. Note that features of the present invention are intentionally designed such that all the new functionality is contained within existing frameworks i.e. it does not violate the already standardised data packet structures using the known protocols referred to above.
The data packets in the MMM data packet train are arranged in decreasing precedence order, hi the example shown in Figure 3, which contains three MMM data packets, the first data packet 42 is one having a payload 62 of the highest orioritv. The second data nacket 44 is one havincr a navlnaH 66 of an

intermediate priority. The third data packet 46 is one having a payload 70 of the
lowest priority.
Precedence values are assigned to each data packet in a descending order, and included in the respective transport protocol header 60, 64, 68, so that the third data packet is discarded during transmission through the packet network infrastructure 16, 18 in preference to the second data packet, and so that the second data packet is discarded during transmission through the packet network infrastructure 16? 18 in preference to the first data packet. Thus, should both the second and third data packets be lost, then the resultant effect upon the most important data is minimised, yet at least some of the least important data also arrives at the destination.
The discarding of data packets may take place at any network node along the path the data takes. If a node is deemed to be congested, then an intelligent process can be used to decide how many data packets must be discarded in order for the congestion to be reduced to an acceptable level. This will take the form of scanning the node buffer, which is currently holding the data to be passed through it. To decide which data packets to discard at a node, the priority levels of the data packets are checked and compared. Starting with the lowest priority first, data packets are discarded until the buffer is sufficiently empty.
Say, for example, there are three data packets in a train, as shown in Figure 3. The data source SI has the highest precedence order, data source S2 has an intermediate precedence level, and data source S3 has the lowest precedence level in the train.
The first data packet has a payload that comprises all the mediums that are necessary to make up the multimedia data, as denoted by data from three different data sources, SI, S2 and S3. As SI is deemed to be the data source with the highest priority or importance value, a large percentage of this data source is allotted to the first data packet in the train, which in turn will have the highest priority of the data packets within the train and hence have the lowest chance of being discarded if there is congestion along the route to the destination.

The payload of the second data packet is partitioned and a lower percentage of data source SI is added to it. This trend continues in the third data packet, where the remaining data from data source SI is allocated. The partitioning is slightly different for data source S2; where in this example approximately a quarter of the first data packet is allocated to S2. The allocation in the subsequent data packets decreases accordingly, although not as rapidly as with SI. As data source S3 has the lowest precedence level, the train is partitioned such that the bulk of the capacity of the third data packet is given to S3.
The scenario depicted in Figure 3 shows the proportion of data source SI in the first data packet 72 to be larger than that in the second data packet 78, which in turn is larger than that in the third data packet 84 i.e. 72 > 78 > 84. The reverse is true for data source S3, with a higher proportion in the third data packet 88 than in the second data packet 82, which in turn is higher than in the first data packet 76 i.e. 76 This partitioning pattern, where decreasing amounts of the highest priority data source are allotted to data packets from the front of the train to the back is just one given example and many other patterns can be formed. The partitioning process is repeated throughout the train in a similar vein for a higher number of data sources and hence a higher possible number of partitions in each data packet. Although, not defined precisely, it is envisaged that the number of precedence levels would be between two and ten in the majority of situations.
Information concerning the type of data and partitioning can be contained in each data packet header 90, 92, 94.
The data packet train length is three here, because the association of the three data packets is necessarily of this length as data from each data source spreads over three data packets. The data from these three sources could alternatively be spread over a higher number of data packets than in this

example, which would give rise to a longer data packet train containing more data packets.
It should be noted that a data packet does not have to contain data from all the data sources. For example, the third data packet 46 could contain only data from the third source S3, and/or the second data packet 44 could contain data from the second source S2 and data from the third source S3 data but not data from the first source S1.
MMM Data Packets having a priori knowledge
During data transmission it may be necessary, due to network congestion, to reduce the size of the payload and allow for a smaller number of data packets to be transmitted to convey the same information. Thus associated with each store and forward buffer is a set of transcoders 102A, 102B, etc. The selection of which transcoder is to be used will be based upon the degree to which the information rate needs to be reduced. The transcoded information is then inserted into the data packet together with the transcoder code of the transcoder used, so that it can be decoded at the destination edge store and forward buffer.
Within the MMM data packet header, there is provided a small data field that can be used to flag the transcoder to be used for a subsequent data packet. This flag provides a form of advance warning data that can be used to prepare a corresponding reverse transcoding process at the second network edge node 18. In one embodiment, the advance warning flag may be inserted into the MMM data packet immediately preceding the data packet in the train in which the differently transcoded data is included. However, it need not be given in the immediately preceding data packet; it could for example be inserted into a packet in the next data packet train or a data packet which is a predetermined number of packets away in the packet sequence. As long as there is some useful relationship with the current data packet, then an advantage can be obtained by insertion of an advance warning flag.
The advance warning process relies on the intelligence in the end points to intelligently fill data packets and pre-organise resources in the receiving end

movement of the subject, with more movement requiring further bandwidth resources to cope with the extra change information between subsequent frames. The International Telecommunication Union (ITU) videoconferencing standard H261 using Quarter Common Intermediate Format (QCTF), which has a refresh rate of 30 frames per second, would be adequate for a mobile phone in a video environment.
The size of the IP data packets is also important as data packetisation delay becomes an issue. For audio data, frames of approximately 60bytes are generated approximately every 20rnsec. This creates an interesting engineering problem, which is beyond the scope of this work. For video, again this depends on the refresh rate, which in turn is content dependant.
The above embodiments are to be understood as illustrative examples of the invention. Further embodiments of the invention are envisaged. It is to be understood that any feature described in relation to any one embodiment may be used alone, or in combination with other features described, and may also be used in combination with one or more features of any other of the embodiments, or any combination of any other of the embodiments. Furthermore, equivalents and modifications not described above may also be employed without departing from the scope of the invention, which is defined in the accompanying claims.Claims
1. A method for transmitting data from a plurality of data processing devices (9,10,11) across a data packet communications network (26) having a congestion control mechanism for reducing the effects of congestion by selectively prioritising data packets, the method comprising the steps of:
receiving data from at least a first data processing device (9) and a second data processing device (10);
constructing a first data packet (42) for carrying data through said network;
constructing a second data packet (44) for carrying data through said network;
attaching prioritisation information (90,92) to at least one of the first and second data packets, the prioritisation information being for use by the congestion control mechanism to prioritise the first data packet in preference to the second data packet; and
transmitting the first and second data packets into said network.
characterised in that the first packet construction process comprises adding data (72, 74) from both the first data' processing device and the second data processing device to the first data packet in controlled amounts, the amount of data from each of the first and second data processing devices added to the first packet being controlled during die first packet construction process; and
the second packet construction process comprises adding data (78, 80) from at least one of the first and second data processing devices to the second data packet.
2. A method according to claim 1, wherein the packet construction process is controlled such that the amount of data from the first data processing device in the first data packet is higher than the amount of data from the second data processing device in the first data packet.

3. A method according to claim 1 or 2, wherein the packet
construction process is controlled such that the amount of data from the second
data processing device in the first data packet, taken as a proportion of the total
amount of data from all data processing device in the first data packet, is lower
than the amount of data from the second data processing device in the second
data packet, taken as a proportion of the total amount of data from all data
processing devices in the second data packet.
4. A method according to claim 1,2 or 3, comprising the steps of:
adding data from the first data processing device to the second data
packet in a controlled amount, the amount of data from the first data processing device added to the second packet being controlled daring the second packet construction process.
5. A method according to claim 4, wherein the packet construction
process is controlled such that the amount of data from the first data processing
device in the second data packet is lower than the amount of data from the
second data processing device in the second data packet,
6. A method according to claim 4 or 5, wherein the packet
construction process is controlled such that the amount of data from the first
data processing device in the first data packet, taken as a proportion of the total
amount of data from all data processing devices in the first data packet, is higher
than the amount of data from the first data processing device in the second data
packet, taken as a proportion of the total amount of data from ail data processing
devices in the second data packet.
7. A method according to any previous claim, comprising the steps
of:
receiving data from a third data processing device (1 i); and adding data (76) from the third data processing device to the first data packet in a controlled amount, the amount of data from the third data processing

device added to the first packet being controlled during the first packet construction process.
8. A method according to claim 7, wherein the first packet
construction process is controlled such that the amount of data from the third
data processing device in the first data packet is lower than the amount of data
from the first data processing device in the first data packet and the amount of
data from the second data processing device in the first data packet.
9. A method according to any previous claim, comprising the steps
of:
constructing a third data packet (46) for carrying data through said network, the process of constructing the third packet comprising adding data (84, 86) from at least the first and second data processing devices to the third data packet;
attaching different prioritisation information (90, 92, 94) to at least two of the first, second and third data packets, the prioritisation information being used by the congestion control mechanism to distinguish between three different levels of prioritisation amongst the three data packets; and
transmitting the third data packet into said network.
10- A method according to any preceding claim, wherein the prioritisation information attached to each data packet is based on delay tolerances, whereby a data packet containing more data from a less delay-tolerant data processing device is given a higher priority and a data packet containing more data from a more delay-tolerant data processing device is given a lower priority.
11. A method according to any preceding claim, wherein the prioritisation information attached to each data packet is based on the importance value of the content of the data packet, whereby a data packet containing data from a more important data processing device is given a higher

priority and a data packet containing data from a less important data processing device is given a lower priority.
12. A method according to any previous claim, for controlling
congestion at a network node in a data packet communications network, the
method comprising the steps of:
receiving at least a first and a second data packet through said network;
prioritising at least one of the first or second data packets in preference to the other, according to prioritisation information contained within at least one of the first and second data packets;
reducing congestion at the node by keeping the data packet with the higher priority level and discarding the other.
13. A method of transmitting data using a plurality of different data
formats across a data packet data communications network (26), the method
comprising the steps of:
selecting a first data format from said plurality of data formats;
adding data to a first data packet, in the first data format;
transmitting the first data packet into the network;
selecting a second, different format from the plurality of data formats;
adding data to a second data packet, in the second data format; and
transmitting the second data packet into the network,
characterised in that before said first data packet is transmitted into the network, advance warning data of the format of the second data packet to be constructed subsequently to said first data packet is added into the first data packet.
14. A method according to claim 13, wherein the first data format is produced by a first transcoder (102A) selected from a plurality of transcoders (102A, 102B, 102C) and the second data format is produced by a different transcoder (103A, 103B) selected from the plurality of transcoders.

15. A method according to claims 13 or 14, whereby the advance
warning data is used to reduce delay by the efficient use of resources, the
method comprising the steps of:
receiving at least a first data packet containing advance warning data; using the advance warning data to prepare for the reception of a second data packet;
receiving said second data packet.
16. A method for transmitting data from a plurality of data
processing devices (9,10,11) across a data packet data communications network
(26), the method comprising the steps of:
receiving data from at least a first data processing device (9) and a second data processing device(10);
constructing data packets (42, 44) for carrying data through said network,
characterised in that the packet construction process comprises adding data (72, 74) from both the first data processing device and the second data processing device to the first data packet (42) in controlled amounts, the amount of data from each of the first and second data processing devices added to the first packet being controlled during the first packet construction process; and
the relative proportions of data from the first and second data processing devices in the data packets are varied in dependence on current conditions of transmission of data through the nfctwork.
17. Apparatus arranged to conduct the method of any preceding claim.



Documents:

3781-CHENP-2006 AMENDED CLAIMS 21-08-2013.pdf

3781-CHENP-2006 AMENDED PAGES OF SPECIFICATION 21-08-2013.pdf

3781-CHENP-2006 EXAMINATION REPORT REPLY RECEIVED 05-10-2012.pdf

3781-CHENP-2006 EXAMINATION REPORT REPLY RECEIVED 21-08-2013.pdf

3781-CHENP-2006 FORM-1 21-08-2013.pdf

3781-CHENP-2006 FORM-3 21-08-2013.pdf

3781-CHENP-2006 FORM-6 21-08-2013.pdf

3781-CHENP-2006 OTHER PATENT DOCUMENT 21-08-2013.pdf

3781-CHENP-2006 POWER OF ATTORNEY 21-08-2013.pdf

3781-CHENP-2006 PRIORITY DOCUMENT 21-08-2013.pdf

3781-CHENP-2006 FORM-18.pdf

3781-chenp-2006-abstract.pdf

3781-chenp-2006-claims.pdf

3781-chenp-2006-correspondnece-others.pdf

3781-chenp-2006-description(complete).pdf

3781-chenp-2006-drawings.pdf

3781-chenp-2006-form 1.pdf

3781-chenp-2006-form 3.pdf

3781-chenp-2006-form 5.pdf

3781-chenp-2006-pct.pdf


Patent Number 257286
Indian Patent Application Number 3781/CHENP/2006
PG Journal Number 39/2013
Publication Date 27-Sep-2013
Grant Date 20-Sep-2013
Date of Filing 12-Oct-2006
Name of Patentee FRANCE TELECOM
Applicant Address 6, PLACE D'ALLERAY, F-75505 PARIS CEDEX 15, FRANCE,
Inventors:
# Inventor's Name Inventor's Address
1 REYNOLDS, PAUL, LURRENCE WATTLES, EASTON, WELLS, SOMERSET BA5 1AT, UNITED KINGDOM
PCT International Classification Number H04L 12/56
PCT International Application Number PCT/GB05/01386
PCT International Filing date 2005-04-11
PCT Conventions:
# PCT Application Number Date of Convention Priority Country
1 0408238.4 2004-04-13 U.K.