[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

WO2004057828A1 - Method of automatically replicating data objects between a mobile device and a server - Google Patents

Method of automatically replicating data objects between a mobile device and a server Download PDF

Info

Publication number
WO2004057828A1
WO2004057828A1 PCT/GB2003/005598 GB0305598W WO2004057828A1 WO 2004057828 A1 WO2004057828 A1 WO 2004057828A1 GB 0305598 W GB0305598 W GB 0305598W WO 2004057828 A1 WO2004057828 A1 WO 2004057828A1
Authority
WO
WIPO (PCT)
Prior art keywords
time
threshold
server
data
replication
Prior art date
Application number
PCT/GB2003/005598
Other languages
French (fr)
Inventor
Thomas Ralph Edwards Greenwell
Stephen Timothy Spence
Mark Colin Stalker
Original Assignee
Cognima Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority to US10/539,863 priority Critical patent/US8156075B2/en
Application filed by Cognima Ltd filed Critical Cognima Ltd
Priority to EP03782665.8A priority patent/EP1576784B1/en
Priority to JP2004561666A priority patent/JP2006511134A/en
Priority to AU2003290306A priority patent/AU2003290306A1/en
Priority to ES03782665.8T priority patent/ES2689498T3/en
Publication of WO2004057828A1 publication Critical patent/WO2004057828A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W4/00Services specially adapted for wireless communication networks; Facilities therefor
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network
    • H04L67/1095Replication or mirroring of data, e.g. scheduling or transport for data synchronisation between network nodes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/50Network services
    • H04L67/60Scheduling or organising the servicing of application requests, e.g. requests for application data transmissions using the analysis and optimisation of the required network resources
    • H04L67/62Establishing a time schedule for servicing the requests
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L69/00Network arrangements, protocols or services independent of the application payload and not provided for in the other groups of this subclass
    • H04L69/30Definitions, standards or architectural aspects of layered protocol stacks
    • H04L69/32Architecture of open systems interconnection [OSI] 7-layer type protocol stacks, e.g. the interfaces between the data link level and the physical level
    • H04L69/322Intralayer communication protocols among peer entities or protocol data unit [PDU] definitions
    • H04L69/329Intralayer communication protocols among peer entities or protocol data unit [PDU] definitions in the application layer [OSI layer 7]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/02Details
    • H04L12/14Charging, metering or billing arrangements for data wireline or wireless communications
    • H04L12/1485Tariff-related aspects

Definitions

  • This invention relates to a method of automatically replicating data objects between a mobile device and a server; data replication is needed, for example, to back up data from the mobile device to the server and to ensure that the mobile device has the most up to date data held on the server.
  • the mobile device and the server are connected over a wireless network, which may comprise a wide area network such as a cellular telephony network.
  • the wireless network may also comprise a short range network, such as an 802.11 network, or a combination of short range, wide area and wire based connections.
  • wireless networks One characteristic feature of wireless networks is the need to use bandwidth efficiently; this applies especially to wide area networks such as cellular networks, but also applies to other kinds of wireless networks, such as those based on 802.11 systems.
  • the present invention is directed to increasing the efficiency of use of all such networks.
  • Cellular telephony network operators currently push data to mobile devices to update those devices with data, such as menus of the names of games that can be download to the mobile telephone.
  • data such as menus of the names of games that can be download to the mobile telephone.
  • a simple example of efficient bandwidth useage is the practice of downloading these menus of games over night to make use of the available bandwidth.
  • a method of automatically replicating data objects between a mobile device and a server, connected together via a wireless network in which the timing of data replication across the network is determined by a network operator applying parameters that make efficient usage of network bandwidth.
  • the present invention is based on the insight that the assumption that data replication must be immediate to be of value is flawed. Instead, it is useful to be able to provide network operators (such as those providing cellular services, or 802.11 services) with the ability to control how data replication services use available bandwidth in order to make the most efficient usage of that bandwidth.
  • network operators such as those providing cellular services, or 802.11 services
  • the parameters applied to a given object may be both time dependent and also relate to how urgently that object needs to be replicated.
  • a change log may list all objects at the device and/or server to be replicated and the parameters then comprise a weight associated with each object that defines how urgently that object needs to be replicated.
  • the parameters may further comprise a threshold that is a function of time, with the weight of each object being locally compared to the threshold at a given time and the outcome of the comparison determining whether the object is sent for replication or not at that time. This combination of weight and threshold gives a flexible way to control the timing of data replication and hence make the best use of bandwidth.
  • the present invention is implemented by Cognima Ltd (London, United Kingdom) to allow mobile network operators to control the timing of data replication in the Cognima ReplicateTM system.
  • the present document assumes a working familiarity of the Cognima ReplicateTM system, which is described in more detail in Appendix 1
  • QoS Quality of Service
  • This IETF document defines QoS very precisely in terms of a number of metrics in the IP layer, whereas the first implementation of Cognima QoS will be applied at the application layer and will not rely upon service-specific configuration of network server parameters.
  • This invention defines a way in which data transmission across a packet- switched wireless network can be intelligently scheduled, to improve efficiency of network bandwidth usage without seriously impairing the user experience.
  • the Cognima Replicate system is designed to replicate user data between a mobile client and a network-based server, with zero user intervention. Both the client and the server recognise when there have been changes to the data which require replication to take place, ensuring the integrity of the distributed object database that is shared by the client and the server. This approach creates an experience for end users of the system, of always having instant access to all their data, without the any need for manual synchronisation. An implication of making replication invisible to the user is that there need be no indication, and no user control, of when the replication is taking place. The client and server devices are in control of when replication should take place, with timing decisions based on dynamic parameters that can be set by the mobile network operator. 1.2 Scheduled Replication And Network Operators
  • Network operators wish to smooth the peaks and troughs of the daily network usage cycle in order to make most efficient use of the bandwidth. This means moving data traffic away from peak times, and where possible moving it into troughs in the cycle. Operators will value the ability to tweak settings that affect when replication occurs, and thereby refine network efficiency.
  • the Cognima technology presents new mental models for users. Replication scheduling models, and their corresponding Service plans, must be simple and consistent to aid user acceptance. Users should be shielded from the details of replication as much as possible. Data should replicate according to users' expectation.
  • the QoS options to the user should be simple. Users will find difficulty in weighing up the relative benefits of a plan that offers Contacts in 2 minutes, Photos in 3 hours, and Banking overnight against one offering Contacts immediateyl , Photos overnight, and Banking in 30 minutes etc, however much these options might fit the Network Operators' demographic research. Users will feel more comfortable choosing a general standard of service - e.g. Basic — and upgrading perhaps one particular service. Users will appreciate the opportunity to temporarily upgrade the replication lag, by service or individual object for a cost. For example, users may want a premium "Send this photo now" option, which would override the default priority with which all other photographs are replicated to the Cognima server.
  • the Network Operator can influence the timing at which all objects are replicated according to the peak and off-peak tariff or periods of high network traffic demand. It is possible for a Network Operator to define a timing profile for each application, against which each device reconciles the replication priorities and time limits of objects in the change-log to determine replication behaviour. The shape of this graph will be determined by a number of factors including the experience of the network operator in monitoring data traffic volumes.
  • the Network Operator can also influence whether other (lower-priority) objects in a change-log at the time of an open data connection should be replicated, once the connection-initiating object has been sent. It is possible to define an opportunism threshold to control this. E.g. several non-urgent items are in the device's change log. The user changes a device setting that immediately initiates a data connection. The NetOp has spe ⁇ fied that any other object in the change-log under 20kB should be replicated using the open connection.
  • a different opportunism threshold should be used if the device is operating in a roamed network, as the cost to the user of initiating connections may outweigh the impact to the NetOp of sending more data at peak periods.
  • the opportunism threshold, along with the other QoS control parameters on the client device, are communicated to the client using the replication framework and are therefore kept automatically in step between the client and the server.
  • the network-based server will be able to determine the cell loading for a given mobile device.
  • the server should be able to signal the client that it may start opportunistic replication. This delivers most benefit in all-IP networks where the client device has a permanently assigned IP address and can therefore be contacted instantly by the server. Otherwise the process of sending a Communications Initiation Request to the client device influences the cell loading, and adds a latency to the system that creates an opportunity for the cell loading to change before replication can start.
  • the Network operator can update the data traffic graph, opportunism threshold and cell loading threshold after deployment. This allows optimisation of the QoS control in the light of experience.
  • the Service Provider has the opportunity to determine a QoS Profile for each Cognima service.
  • This profile contains sets of replication time limits within which Cognima objects created by that service are intended to be replicated.
  • the actual time at which replication is attempted within the time limits is determined by a number of elements in the system including the state of the Cognima client software and the client device, and network factors such as data tariff peak/off peak times, cell loading and so on.
  • the Service Provider can set a zero time limit - i.e. request immediate replication for objects of a given type. If replication is attempted and fails, the object remains scheduled for replication but is subject to the existing backing-off behaviour.
  • E.g. Provider specifies that replication of new contacts to the Server should be attempted within a 2 hour time limit. If a contact is created during a network trough or while the cell loading is very low, the device may send the contact immediately. If the timing coincides with a network peak or the cell loading is high, the device can wait for up to 2 hours for conditions to change, but after this time it must attempt replication regardless.
  • the QoS Profile defines the following factors as parameters determining replication timing:
  • Object type e.g. Contact, Photo, DCC directory, DCC usage log etc
  • How object was created e.g. new contact on handset, edit to contact created on portal etc.
  • Direction of travel e.g. edits to contact on portal go immediately, edits to contacts on phone go within 2 hours
  • ⁇ Handset is on home/roamed network (e.g. photo replicates immediately at home, but within 12 hours if roaming)
  • the time limit for replication is assigned to an object by a QoS Profile based on its time of creation. However it is also possible to change a limit already assigned according to subsequent events such as the memory available on device changing, or if the handset roams to another network. This creates a need for occasional recalculation of the weights of the items in the changelog.
  • the Service Provider is able to define a shelf life for objects in the QoS Profile. If an object reaches its shelf life while it is still in the change log, the object should be deleted.
  • the Service Provider specifies that a weather update in the DCC service has a shelf-life of 24 hours. If the item hasn't replicated to a handset within this time, it is deleted from the change-log
  • the Service Provider can define an object as over-writeable. If a new object enters the change log that replaces an earlier version still in the change-log, the earlier version is deleted. In this situation, the Service Provider can determine whether the new object should adopt the timing of the object that has been overwritten, or should enter the system with a new replication time limit.
  • the default setting is that new objects adopt the timing characteristics of the objects they replace.
  • the Service Provider specifies that a objects of the weather update class are over- writeable, and that new entries should adopt the timing of the ones that get overwritten. This ' will mean that an old forecast message gets overwritten by a newer one, but the newer one is not held back from replication.
  • Service Providers can provide different classes of service within the Profile set.
  • One way of dong this is by defining individual QoS Profiles for each class of service.
  • Contacts in a Gold service are replicated immediateyl after creation, but under the Silver service on the same network, replication may take up to 2 hours.
  • a QoS tariff applies to an individual Cognima service, although it may be presented to the subscriber as a bundle of services at a particular price point.
  • a QoS Profile can be changed by the Service Provider once it is deployed; changes to the Profile are replicated to the relevant client devices so that the understanding of each QoS profile is common between all entities in the system. 2.3 Notes on deriving a QoS Profile.
  • QoS is implemented as an increment to the functionality of both the client and the server. In particular it requires a modification to the behaviour of the change log, and introduces a requirement to recalculate certain attributes of queued change log entries. Replication can then occur as a result of the outcome of this recalculation.
  • the algorithm is made up of several components.
  • a changelog item weight This weight indicates how urgently a changelog item needs to be sent; the heavier the weight, the more urgent the item.
  • a changelog threshold Any items that have a weight equal to or greater than the threshold need to be sent immediately. When a connection is made all items with weight greater than the threshold minus a delta are sent. The delta represents opportunism.
  • weight and threshold can vary between 0 - 100.
  • the weight of an item that absolutely has to be sent right now is 100; a changelog threshold of zero indicates that any entries in the changelog can be replicated right away.
  • Both weight and threshold can vary over the course of a day. There will be predictable variation and also dynamic variation. Some examples will clarify this.
  • the straight line at weight value 40 shows that the weight is constant over time.
  • the weight of an item that has to go within a certain time is shown at Figure 2.
  • the weight of the item starts at a relatively low value, indicating low priority, and then it jumps as we reach the time limit - the new value of 100 will force the client to attempt replication of the object at time TI.
  • the threshold will also have a graph over time, the following Figure 3 graph shows an example of how a daily threshold cycle might appear, with a high threshold to guard against low-value traffic during peak hours (e.g. after 9.00 hours for several hours), and a lower threshold when the data networks are traditionally quieter (between 24.00 hours and 6.00 hours, the threshold is zero).
  • the example effectively shows a daily cycle split into three tariff bands, perhaps named off-peak, standard and peak, with the replication threshold set appropriately for each band. Note that there is a fourth band just after midnight, where the threshold drops to zero - this is introduced to ensure that all changelogs are emptied once per day during the quietest period. This zero threshold period could be defined once per week or at some other interval defined by the network operator, but is recommended to ensure that the defined QoS profile does not prevent some objects from being replicated at all. As for the other bands in the profile, the off-peak period extends through the late evening and night, and represents the times at which the data network can expected to experience low traffic; replicating large objects during this time will allow ' the mobile network operator to make best use of the limited bandwidth available during peak times.
  • the zero-threshold period can be adjusted by the network operator in the light of experience; finer adjustments can be made for different user groups, per-application or even per-user thereby staggering the off-peak network usage and ensuring that the full user base for a given server does not attempt to connect to the Cognima Server at the same time.
  • threshold for example from cell loading or to support a marketing drive promoting a new service (during which time it may be preferable to allow basic tariff data onto the network during peak hours to encourage uptake). If it is possible for the device to detect that its cell is not busy it could drop its threshold a bit which might trigger some replication.
  • the core of the algorithm is to calculate the threshold graph and the weight graph of every item in the change log. If the current weight of any item is greater than the current threshold then a connection is established. Otherwise the next time that any item will exceed the threshold is deduced (ie when the graphs intersect) and a timer is set for this interval. Because both the weights and the threshold can be dynamic there are several events that can trigger a recalculation:
  • the server may push a new threshold value to the client. This is actually a special case of the previous event since the QoS object on the client is controlled through replication in the normal way, meaning that a new threshold value will be delivered to the client by placing the change in the device's changelog queue on the server - this changelog entry must have a weight of 100 to force immediate replication, and the resulting change of threshold may trigger or delay replication of other entries already in the queue.
  • a timer expires this will usually be the timer indicating the point at which the weight of an existing entry in the changelog is due to exceed the current changelog threshold. • The cell (or network) loading changes
  • the client software After each refresh of the changelog, the client software also calculates the time interval to the next intersection between a weight graph and the threshold graph; this is so that scheduled replication can take place as required even if there are no further changes before that scheduled time.
  • the 'next event' for the change log will either be as predicted by this intersection calculation, or will be created by some external event that places a new entry in the change log (which may of course force a refresh of the values of all weights and the threshold). Note that the 'next connection' event may have a value of 'never' if the changelog is empty, unless the active QoS Profile has a zero-threshold period as in the example above.
  • Bands can be defined as deltas from a start time or against the system clock (absolute).
  • a weight graph for a class which should be scheduled to go within 2 hours of object creation could be described by a pair of bands: The following table could describe the QoS weight profile for the Contacts class.
  • a threshold graph which describes the graph in the example above would be represented as follows:
  • the client and server carry QoS objects which encapsulate the above tables and influence the scheduling of replication.
  • a user's device will hold a QoS object for each data class installed on the device, plus a single threshold object representing the whole system.
  • Each QoS object holds an array of bands to describe a single weight graph. The intersection points of these graphs determine when objects of a given type will be replicated - these intersection points must be calculated by determining the weight of each object and the threshold value for the system at a given time. 3.4 Changelog item weight calculation
  • the weight of an item in the change log has to be calculated.
  • the QoS object for the item's class is retrieved from the data store.
  • the banding structure is examined and a weight is looked up from the weight graph (either by calculating current time — created time in the case of delta band type, or by comparing system time in the case of absolute band type).
  • the oversize weight is used — note that this could be higher or lower than the default weight for the class.
  • the object weight can be overridden by a user request (effectively a 'send now' button) which sets the weight to 100.
  • the current changelog threshold weight can be extracted from the threshold QoS object using the current system time. This value can then be modified by dynamic variables — for example if the device can detect roaming status then this can influence the threshold. Generally the threshold will be higher when roaming, to reflect the fact that replication will be more expensive; it is also possible to specify a lower limit for the threshold in a given band, effectively preventing very low priority objects from replicating at all before the device returns to its home network. In deployments where the mobile terminal can be aware of local cell loading conditions, then the cell loading can be used as a factor in adjusting the current threshold value: if the traffic loading of the local cell is below some value (e.g. 70%) then the threshold can be reduced. If the cell loading is above some value (e.g. 95%) then the threshold can be increased.
  • some value e.g. 70%
  • the QoS object creates timer events which represent the times of day at which the threshold is known to change, reflecting the shape of the QoS Profile as defined by the network operator. As each of these timer events fires off, the QoS threshold will be assigned a new value and the ensuing weighting recalculation will allow objects with the correct weight to be replicated.
  • the QoS threshold may drop due to a move from one time-based tariff to another
  • Cognima has developed a data replication technology that directly addresses the need for Mobile Service
  • MSPs Mobile Subscribers
  • Network Operators to increase consumer adoption of data services, encourage greater loyalty from their valuable customers, and differentiate their services from the competition.
  • Cognima's data replication solution addresses these issues by: • Increasing adoption by making data services compelling and effortless to use.
  • the Cognima's data replication framework enables a Mobile Service Provider to build compelling services for consumer markets.
  • the MSP hosts a Cognima Server at its data centre.
  • the server comprises an Oracle database plus
  • the Cognima framework replicates data entered in a mobile phone automatically
  • Always-available means that the personal information is accessible on whichever device or handset the subscriber is carrying, whether currently connected to the network or not since the user can always access personal information stored locally on the device). Users can also edit and manage their personal data directly on the server via a web interface - the Virtual Phone.
  • Cognima provides an ideal framework for implementing mass-market consumer data services based on the following key benefits:
  • the MSP can control when replication takes place, and the Quality of Service (QoS) delivered.
  • QoS Quality of Service
  • the MSP can offer low-cost, subscription-based services with relatively poor QoS without sacrificing user experience - e.g. data replication can happen overnight for non-urgent data services such as bank statements, yet still be satisfactory to users.
  • Overnight data replication uses otherwise underused bandwidth and is hence far cheaper than peak time data replication.
  • Urgent data replication e.g. presence information
  • efficient use of phone memory & processor power allows Cognima client software to be cost-effectively installed in even the cheapest mass-market phones. 3. Benefits to the Mobile Service Provider
  • Cognima presents a MSP with a means to generate new data revenues, reduce churn, and to differentiate its services from those of its competitors.
  • Cognima enables rapid introduction of compelling and affordable new mobile data services. • Cognima delivers a compelling user experience for new services in low-end phones using only spare network capacity. This is affordable and scalable for the network operator, allowing the MSP to offer understandable and predictable pricing for mass-market subscribers.
  • Cognima's client software can be installed as a flash memory upgrade, endowing today's mass-market handsets with smart-phone-like capabilities. New software applications can be downloaded over the air to existing Cognima-enabled handsets, allowing MSPs to roll out new data services without waiting for new devices to support them.
  • Cognima services act as a significant barrier to churn. For example, a subscriber who stores their personal information securely at their MSP's Cognima Server can buy a new phone and immediately retrieve all personal information to their new device. All this personal information may be lost if they decide to take out a subscription with a different service provider.
  • Cognima gives MSPs the ability to implement services on the handset, and thereby to regain control of their subscribers' user experience. Most importantly, Cognima allows this without sacrificing interoperability; support for industry standards is achieved through straightforward integration with the Cognima Server. The net result is that the MSP's position in the value chain is strengthened versus the powerful brands of handset manufacturers and content providers.
  • the objectives of the Cognima software are:
  • Client devices hold a copy of the user's data in a database on the client device. The user can access this data whether or not he has a network connection and therefore always has instant access.
  • the client device connects periodically to a Cognima Server on the wireless network, to send up the changes from the Change-Log and receive new data. This separates the act of changing data from the need to connect to the network (i.e. push is not continuous in a preferred implementation).
  • the Cognima Server updates its own database with data changes received from the client device, and populates Change-Logs for any other devices the user owns. When these devices next connect, they will receive the changes and thus the devices are kept in sync, each with a copy of the same data.
  • the Cognima Server contains a web server which allows the user to examine directly using a web browser the copy of the data held in the Cognima Server database, and make changes to it as he would on a client device.
  • the Cognima contains a web server which allows the user to examine directly using a web browser the copy of the data held in the Cognima Server database, and make changes to it as he would on a client device.
  • the client device can effectively ask the Cognima Server to send a message as an SMS or an email or a fax by setting a few flags in a message object and the Cognima Server contains the functionality to communicate with email servers, SMS servers and fax machines. This can be extended to servers holding ringtones, banking details, games etc. It is easier and cheaper to build the software on the Cognima Server to talk to these other servers, than it would be to build the software on the client device.
  • All users in a Cognima network are assigned a user id. This id is unique to the network — i.e. provided by a given network operator. All users have a Cognima address which is a combination of their user id and Cognima Server URL. This is unique in the world. Each device which belongs to a user is assigned a device id. The device id is unique to the user. This is only 8 bits so a user can have a maximum of 253 devices (id 254 is reserved for the web, id 255 is spare, id 0 is invalid). All user data is classified into classes (contacts class, messages class, bank transactions class etc) and the classes are assigned a class id which is unique in the world. Class id '12' refers to a contact, for example.
  • An instance of a class is an object, which is assigned an object id unique to the user, e.g. a contacts class object might be the contact for "John Smith".
  • the object id is generated by concatenating the device id of the device which created the object with a monotonic increasing count which increases over the life of the device. So each device can create a maximum of 16777215 objects (if we encountered this limit we could reset the device id).
  • Classes are defined by the properties which constitute them.
  • a class is essentially an array of properties. Each property in the class has a property id which is unique to the class (and is actually just the array position of the property in the property array, starting from zero). 5.1.2 Creating objects
  • An object is created on a device. It is assigned an object id and saved to the device database. A copy is also saved into a Change-Log. When the device next connects to the Cognima Server the entry in the Change-Log is sent up. The Cognima Server saves the object to its database (recording the system time), does any class specific processing that may be required (such as generating and sending an email) and adds entries to Change-Logs for any other devices that the user may own which have declared interest in the class. (The entries should be for the correct version of the class on the device).
  • An object may also be created on the web portal.
  • the object id is generated
  • An object may also be created by a server application (e.g. a messaging module might receive an email from which it creates a message object).
  • the object id is generated (using device id of 254 as described above) and processed identically to the device.
  • One or more properties of an existing object are modified on a device.
  • the changes are saved to the device database.
  • Each changed property is used to generate an entry in the device Change-Log. These are sent up to the Cognima Server.
  • the Cognima Server saves the changes to its database (recording the new ast changed' time for the property), does any required class specific processing and adds entries to Change-Logs for other devices which belong to the user, have declared the class and have a version of the class which contains the property.
  • the update is also placed on the Change- Log for the device that originated the change. This may seem strange but is required to cope with the following scenario:
  • a user has 2 devices A and B. He updates property 7 on A offline at 5pm and updates it on B offline at 6pm. He connects to the network with A first. The value of 7 on A gets put in the Change-Log to be sent to B. Later B connects. Its value of 7 is more recent so the value of 7 on B is sent to A, but B getsA's value. Replicating the value of 7 back to B fixes this.
  • An object is deleted on the device. It is removed from the device database and an entry is put on the Change-Log listing the class id and object id. The entry is sent up to the Cognima Server. If the time of the deletion is later than the last updated time of the object, then the Cognima Server marks the object as deleted in its database, does any class specific processing and adds the entry to other devices that belong to the user and have declared the class.
  • the deleted object is viewable in the web portal a manner that makes its deleted status clear.
  • the user can select the object for un-deletion.
  • the deletion mark is removed from the object in the Cognima Server database and entries to refresh the object are placed on the Change-Logs for all devices that belong to the user and have declared the class.
  • Each property has a type. There are currently 9 permitted property types:
  • a CogString is a character count followed by the characters. If the string is ASCII then the space taken up by the string will be (4 + char count) bytes. If the string is Unicode then the space taken up will be (4 + (char count * 2)) bytes.
  • a CogTypedString is a CogString preceded by a type (4 byte unsigned integer). The only use of a typed string so far is a Contact Point. The type identifies the type of contact point (e.g. email address, home phone) and the string holds the address (e.g. bob@xxx.yyy, 01233556677).
  • a CogBlob is a length in bytes followed by that number of bytes. It can be used to store any binary data.
  • a CogArray is passed around as a 4 byte unsigned integer 'type' followed by two blobs.
  • the 'type' indicates the type of elements held in the array.
  • the first blob is an index blob: it holds a sequence of offsets (4 byte unsigned integers) into the second blob.
  • the second blob is the data blob which holds the elements of the array as a sequence of binary lumps. Elements can be extracted from the data blob by counting along the index blob to get the offset of the start of the element in the data blob.
  • This is the stream structure of the CogArray as it is passed around. Inside a particular system it may appear as a conventional vector (i.e. already parsed).
  • CogArray The only implemented example of a CogArray is the MessageAddress. Each element of the MessageAddress is an AddressPair.
  • An AddressPair is a contact id (object id of a contact object) followed by a Contact Point.
  • Some of the properties can be made "smart". This means they can be parameterised for a specific device to sculpt the data in the property for the characteristics of the device.
  • the parameters are two 4 byte unsigned integers, one is a smart type and the other is a max size.
  • the property which holds the body text of a message might be parameterised to smart type kCogPlainText and max size 100 on a cheap phone with limited memory, but parameterised to be smart type kCogRichText and max size 1000 on a PDA with more memory.
  • the parameters are stored by the Cognima Server when the application is added to the device.
  • new objects or updates for that class are placed in the Cognima Server Change-Log for that device they are processed according to the, smart parameters. This might involve, for example, truncating text, converting Unicode text to narrow text or converting image formats.
  • Class versions We have the concept of a class version which is defined by a 4 byte unsigned integer. A new class version may add properties to the end of the old class, but it may not change or remove existing properties, or insert new properties between existing properties. This should allow interoperability between versions. Class definitions with different smart property parameters are not different versions.
  • Class metadata is essentially an array of property metadata.
  • Property metadata is a property id, a property type, a smart type and a max size.
  • User data is transferred as a stream with no formatting information other than a class id. This stream is parsed by looking up the class metadata.
  • Client devices declare to the Cognima Server the classes that they support. This enables the device to subsequently send up only raw user data (with a header containing class id, object id and a few other things) and hence minimises bandwidth requirements. This can be contrasted with, for example, XML reliant systems that are far more bandwidth hungry.
  • the client device class declarations also contain the smart property parameters so that the Cognima Server can sculpt the data for the device. It is worth emphasising that the meaning of a property is hard coded into an application.
  • the class metadata states that property 2 in class 7 is a string with max length 30 characters. It is the code in the application that interprets property 2 in class 7 as the name of a football team.
  • a key part of the design was to not require times of modification to be kept for each property of an object on the client device as updating these on constrained client devices is slow and keeping a last modified time for each property in an object would take a lot of space.
  • Class Metadata definition tells the CRE (Cognima recognition engine) what type a given property is and allows it to pack and unpack an object or a property for transmission over a data link.
  • CRE Recognition Engine
  • the purpose of the ChangeLog is to record any changes that have occurred since the client device last connected to the Cognima Server (or the Cognima Server to the client device).
  • Cognima APIs applications connect to the CRE and can cause objects to be created or deleted, or a property in an object to be changed. These changes are added to a Change-Log on the local device as they are made together with the time the change was made. Objects are given unique identifiers when they are created so that a given object can always be identified.
  • a Cognima Server receives ChangeLog items from a client device:
  • a Cognima Server When a Cognima Server receives a new object from a client device it is added to the Cognima Server database and also added to all the Change-Logs of the client devices registered to that user, apart from the Change-Log of the machine that sent the new object in the first place.
  • a Cognima Server When a Cognima Server receives a property change it compares the time of the change to the current time held for that property on the Cognima Server. If the time of the property change is later than that held on the Cognima Server the property value is changed in the server database and this change is also added all the Change-Logs of the client devices registered to that user - including the one of the machine that sent in property change (in case another object update has been sent to that machine in the meantime). If the property change was not later than the one held on the Cognima Server no change is made as the stored property value is more recent - but the value is added to the list of old property values on the Cognima Server so that a user can retrieve it later if required. When times are compared the Time Management approach described in Section 6.2.below is used.
  • a device When a device first connects to a Cognima Server it will be sent all class metadata definitions and then all the objects in the database for that user.
  • the Deletion messages generally just mark an object for deletion. Actual removal of the object from the database may occur later on once all objects referring to that object have also been deleted.
  • Operating systems are fundamentally little endian or big endian which is a choice of the byte order in which numbers and strings are stored. If two computers which have different endian-ness have to communicate then one of the computers will have to switch the endian-ness of its data packets.
  • the Cognima client software uses the same endian-ness as the host client device.
  • the Cognima Server has to determine the endian-ness of the client device (it uses a reference value in the first packet of data from the client) and then convert the subsequent incoming data if necessary to maintain consistent endian-ness in the Cognima Server.
  • the Cognima Server also has to convert any outgoing data it sends back to the client device.
  • the logon of a device contains the current device time.
  • the Cognima Server should be able to compensate for the latency of the network and compare the login time with its own system time. This will give it a delta between the device time and the Cognima Server time. This delta can be applied to further times sent up by the device in that session.
  • the Cognima Server can compare deltas in successive sessions from a device to determine clock 'creep' on the device or changes of time zone: it cannot be assumed that all the client devices in the system have clocks that are well synchronised to each other: • Clock times drift on devices depending on the device's clock accuracy.
  • the server will be responsible for adjusting times used by the client device to GMT when comparisons are made on the Server, and from GMT to the equivalent time for the client device when messages are sent from the Cognima Server to the client device.
  • the client device will tag all the items in the ChangeLog with times obtained from the local clock - as far as the client device is concerned it only ever deals in time based on the client device's own clock.
  • the table below shows a pattern of events with a client device connecting to a Cognima Server.
  • the Client device's time is 5 minutes slower that the Cognima Server and is loosing a minute every hour (an extreme case to show the point). Also to show the point we will assume that from 09:00 to 12:00 the user is on a plane and out of contact with the Cognima Server so it does not connect during this time:
  • the Cognima Server In order to work out if the property changes were made before or after the time stored on the Cognima Server the times X and Y need to be worked out. From the information above the Cognima Server knows that when the client last connected it was around 3 hours ago and at that point the time difference was 5 minutes whereas now it is 8 minutes. Thus, assuming the clock drift happens linearly, the Cognima Server can work out that the device is 5 minutes behind GMT and that the clock is drifting back a minute every hour.
  • Property B needs to be adjusted to 11:07 - the 5 minutes initial drift plus 2 minutes since two hours elapsed from 09:00 to 11:00 when the property was changed.
  • the delta to the time between the client device time and GMT may be minutes, but the drift will be in the order of fractions of seconds per hour.
  • users can also change the time on the client device. They may do this to reset the time to the correct local time (we can give the user the option to have this happen automatically but some users may want to keep their own control of their client device time — e.g. they like to have the clock set 5 minutes fast). They may also make adjustments to reflect a change of local time (i.e. daylight savings or changing timezone). The goal is that the user can change the clock on the device to any time that suits the user and the device simply takes account of this.
  • the client device When the user makes a change to the client device time most operating systems will report this change (for systems that don't do this the time can be polled say every minute to check for such a change). On detecting a change in time the client device will work out the delta between the new time and the time as it was before the change. For example this may be a change of plus one hour as a user moves timezone. The client device stores this time difference as the Adjust Time which it saves for the next connection to the Cognima Server. The client device also goes through every entry in the ChangeLog and updates all times in the log by Adjust Time. This ensures that the entries in the ChangeLog are always relative to the local time on the client device.
  • the client device When the client device next connects to the Cognima Server the client device sends at logon the stored Adjust Time — i.e. the amount by which the client device clock has been adjusted backwards or forwards since the last connection.
  • the Cognima Server can then remove this amount from the time from the delta to GMT and drift calculation.
  • An application will use one or more classes to hold user data.
  • the definition of the class is hard coded into the application.
  • the version of the class is coordinated by releases of the application.
  • a statistics application uses a Footballer class to hold data about footballers.
  • the application starts on a client device for the first time, it inquires from the device what version of the Footballer class the device already holds. If the version on the device is the same as the version that the application has been hard coded to use then nothing more need be done.
  • the Cognima Server therefore maintains a list of versions of classes used on all devices.
  • the web portal pages will be the equivalent of the hard-coded device application.
  • the web can extract objects from the database according to the latest version of the class, and if there are more properties than it was hard coded to expect it can ignore them. Therefore the web does not need to declare class versions.
  • the Cognima Server maintains Change-Logs for all devices listing changes that will be sent to the devices when the devices next connect. There will be optimisations that can be made to the Change-Logs, for example: • If >2 updates to the same property are queued in the Change-Log then only the last need be kept.
  • the space available on a client device to hold user data will typically be orders of magnitude less than the space available on the server.
  • the device needs to hold a subset of data and the user should have to do as little work as possible to maintain this subset. ghosting and withdrawal are tools to aid this.
  • a class definition may include flagging certain properties as 'ghostable'. This means that if the object is ghosted those properties will be nulled, freeing room on the client device. ghosting is done automatically on the device.
  • the decision about ' which objects to ghost is made by following a 'ghosting rule' and applying the rule whenever an object is created or updated.
  • the rule defines the maximum number of a selection of objects. When the maximum is exceeded the objects in the selection at the bottom of a sort order are ghosted.
  • the class might be messages
  • the selection might be messages in the inbox
  • the sort order might be by date/time and the maximum number might be 50. If there are 50 messages in the inbox and a new message arrives, the oldest message in the inbox is ghosted.. ghosting may remove the message body but leave enough- header information for the message to be recognised. Withdrawal (also known in the past as auto-deletion and removal) is similar to ghosting but works by removing the entire object, not just part of it.
  • a request is passed from the client to the Cognima Server for the object to be resurrected.
  • a refresh of the object is sent down to the device and the object is put back to normal.
  • a pinned object is never ghosted or removed.
  • Pinning can be chosen by the user, or it can happen automatically. For example, an object that is resurrected is automatically pinned.
  • a set of one or more Cognima addresses is attached to the object which is to be shared.
  • the object can be set to read-only (so the people you share it with cannot modify it).
  • the Cognima Server receives the new object (or receives an update to it) from the web or a client device it replicates it as normal.
  • the Cognima Server of the sharee receives the object. If it is a new object it assigns a new object id (keeping note of the originator id). If it is an update it finds the object using the originator id. If the sharee is allowed to update the object, the update can be replicated back to the object owner using the originator id.
  • a Cognima application does not read in all the data it deals with from a database. Instead it creates a selection — a subset of the data which it is currendy interested in. In general this selection matches the data that is currendy being displayed on the devices screen. Thus only the data currently being used by the application is held in memory - saving a lot of memory space.
  • the application can also specify how the data is sorted and if only a small window on the sorted list of data is required (known as a view).
  • This approach is similar to the screen re-paint approach used to redraw graphics screens on Windowing systems.
  • the application that is responsible for that bit of screen is called to repaint the screen.
  • Example A client device may have a contacts application running on it - this device replicates data with a Cognima Server connected to other client devices also running contacts applications.
  • a class of object is defined for a Contact that contains names and phone numbers and these are replicated to all the devices of a given user.
  • An application on one device may have a display that shows all contacts by beginning letter - for example the interface allows the user to press a D button to show all the names beginning with D.
  • This application will set up a selection that contains objects: • Where the class is defined as Contacts
  • Sorting Data in selections generally needs to be sorted - often so that when displayed users can see data in a logical format.
  • the sorting order can be specified: the properties to sort on, in what order and what sorting algorithms to use.
  • a view provides this functionality by specifying the number of items of data the selection wants to deal with and the number of the first item of data out of the complete list of data the application wants to appear in the selection.
  • Cognima has made some efficiency optimisations in how the data is transferred between the Cognima server and client application — when multiple data changes are made the data is sent in blocks and then the application informed that the changes are complete so that the application only needs to update its user interface once.
  • ObjectAddedO ObjectUpdatedO
  • ObjectRemovedO The methods ObjectAddedO, ObjectUpdatedO and ObjectRemovedO are called by the framework whenever respectively an object is added, updated or removed.
  • the constructor simply calls the default COdbSelect constructor.
  • the ObjectAddedO, UpdatedO and RemovedO methods print out what change was made and then call ListContactsO to show what the current contents of the list is.
  • the ListContactsO shows how the current list of object held by the selection can be accessed.
  • the current list of pointers to objects is held in a container class called iResult - this can then be accessed by normal container class integrators. In this case we simply go through the list and print all the objects in the list.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Computer Security & Cryptography (AREA)
  • Mobile Radio Communication Systems (AREA)
  • Data Exchanges In Wide-Area Networks (AREA)
  • Telephonic Communication Services (AREA)

Abstract

Network operators can control how data replication services use available bandwidth, in order to make the most efficient usage of that bandwidth, using parameters applied to a data object to be replicated. The parameters may be both time dependent and also relate to how urgently that object needs to be replicated. A change log lists all objects at the device and/or server to be replicated and the parameters then comprise a weight associated with each object that defines how urgently that object needs to be replicated; the weight of each object is then locally compared to a threshold at a given time and the outcome of the comparison determines whether the object is sent for replication or not at that time. This combination of weight and threshold gives a flexible way to control the timing of data replication and hence make the best use of bandwidth.

Description

METHOD OF AUTOMATICALLY REPLICATING DATA OBJECTS BETWEEN A MOBILE DEVICE AND A SERVER
FIELD OF THE INVENTION
This invention relates to a method of automatically replicating data objects between a mobile device and a server; data replication is needed, for example, to back up data from the mobile device to the server and to ensure that the mobile device has the most up to date data held on the server. The mobile device and the server are connected over a wireless network, which may comprise a wide area network such as a cellular telephony network. The wireless network may also comprise a short range network, such as an 802.11 network, or a combination of short range, wide area and wire based connections.
DESCRIPTION OF THE PRIOR ART
Data replication between mobile devices (such as mobile telephones, smart phones, communicators and other kinds of wireless information device) and servers has attracted considerable attention. Reference may be made to PCT/GB2002/005308 and PCT/GB2002/005311 (to the present applicant), the contents of which are incorporated by reference.
One characteristic feature of wireless networks is the need to use bandwidth efficiently; this applies especially to wide area networks such as cellular networks, but also applies to other kinds of wireless networks, such as those based on 802.11 systems. The present invention is directed to increasing the efficiency of use of all such networks.
Cellular telephony network operators currently push data to mobile devices to update those devices with data, such as menus of the names of games that can be download to the mobile telephone. A simple example of efficient bandwidth useage is the practice of downloading these menus of games over night to make use of the available bandwidth.
To date, designers of data replication systems (as opposed to simple one-way push updating systems) have not been preoccupied with making efficient use of network bandwidth. This bias arises because the assumption behind most data replication systems is the need for immediate replication, irrespective of the impact on bandwidth.
SUMMARY OF THE PRESENT INVENTION
In a first aspect, there is a method of automatically replicating data objects between a mobile device and a server, connected together via a wireless network, in which the timing of data replication across the network is determined by a network operator applying parameters that make efficient usage of network bandwidth.
Hence, the present invention is based on the insight that the assumption that data replication must be immediate to be of value is flawed. Instead, it is useful to be able to provide network operators (such as those providing cellular services, or 802.11 services) with the ability to control how data replication services use available bandwidth in order to make the most efficient usage of that bandwidth.
The parameters applied to a given object may be both time dependent and also relate to how urgently that object needs to be replicated. For example, a change log may list all objects at the device and/or server to be replicated and the parameters then comprise a weight associated with each object that defines how urgently that object needs to be replicated. The parameters may further comprise a threshold that is a function of time, with the weight of each object being locally compared to the threshold at a given time and the outcome of the comparison determining whether the object is sent for replication or not at that time. This combination of weight and threshold gives a flexible way to control the timing of data replication and hence make the best use of bandwidth.
Further aspects and details are defined in the appended Claims. BRIEF DESCRIPTION OF THE DRAWINGS
The present invention will be described with reference to the accompanying drawings, which show graphs of how the parameters used to control data replication can vary over time.
DETAILED DESCRIPTION
The present invention is implemented by Cognima Ltd (London, United Kingdom) to allow mobile network operators to control the timing of data replication in the Cognima Replicate™ system. The present document assumes a working familiarity of the Cognima Replicate™ system, which is described in more detail in Appendix 1
It should be noted that the term Quality of Service, or QoS, is used throughout the present document, but unless stated in context this is not related to the technical meaning oi QoS in the context of RFC2475. This IETF document defines QoS very precisely in terms of a number of metrics in the IP layer, whereas the first implementation of Cognima QoS will be applied at the application layer and will not rely upon service-specific configuration of network server parameters.
1. Scheduled Replication of Data
This invention defines a way in which data transmission across a packet- switched wireless network can be intelligently scheduled, to improve efficiency of network bandwidth usage without seriously impairing the user experience.
1.1 Cognima Replicate
The Cognima Replicate system is designed to replicate user data between a mobile client and a network-based server, with zero user intervention. Both the client and the server recognise when there have been changes to the data which require replication to take place, ensuring the integrity of the distributed object database that is shared by the client and the server. This approach creates an experience for end users of the system, of always having instant access to all their data, without the any need for manual synchronisation. An implication of making replication invisible to the user is that there need be no indication, and no user control, of when the replication is taking place. The client and server devices are in control of when replication should take place, with timing decisions based on dynamic parameters that can be set by the mobile network operator. 1.2 Scheduled Replication And Network Operators
Network operators wish to smooth the peaks and troughs of the daily network usage cycle in order to make most efficient use of the bandwidth. This means moving data traffic away from peak times, and where possible moving it into troughs in the cycle. Operators will value the ability to tweak settings that affect when replication occurs, and thereby refine network efficiency.
Operators will also wish to offer services of different QoS and cost levels (usually expressed as a choice of service level bundles with their existing data tariff) to address the varying demands of customers. This is enabled by providing the network operators with the opportunity to set dynamically, the parameters which define replication QoS from the users' perspective.
Most importantly the services that Operators offer should provide a compelling user experience to their customers.
1.3 Scheduled Replication And User Expectations
The Cognima technology presents new mental models for users. Replication scheduling models, and their corresponding Service plans, must be simple and consistent to aid user acceptance. Users should be shielded from the details of replication as much as possible. Data should replicate according to users' expectation.
Users want to be able to choose one QoS level from a range of options; otherwise they feel that they are either paying too much or not getting good service. The QoS options to the user should be simple. Users will find difficulty in weighing up the relative benefits of a plan that offers Contacts in 2 minutes, Photos in 3 hours, and Banking overnight against one offering Contacts immediateyl , Photos overnight, and Banking in 30 minutes etc, however much these options might fit the Network Operators' demographic research. Users will feel more comfortable choosing a general standard of service - e.g. Basic — and upgrading perhaps one particular service. Users will appreciate the opportunity to temporarily upgrade the replication lag, by service or individual object for a cost. For example, users may want a premium "Send this photo now" option, which would override the default priority with which all other photographs are replicated to the Cognima server.
1.4 Scheduled Replication and the Cognima Architecture
From an engineering point of view, solutions should be based on simple models rather than complex rule-sets. Rules that depend on time-outs (e.g. each object should have its own time limit that starts counting down when it enters the Change-Log) will severely affect performance. Likewise a solution requiring the CRE to poll the Change-Log every x seconds will also reduce performance.
2. QoS Implementation 2.1 The QoS Profile
We introduce the concept of a QoS Profile which determines when objects of a given type will be replicated for a given user.
The Network Operator can influence the timing at which all objects are replicated according to the peak and off-peak tariff or periods of high network traffic demand. It is possible for a Network Operator to define a timing profile for each application, against which each device reconciles the replication priorities and time limits of objects in the change-log to determine replication behaviour. The shape of this graph will be determined by a number of factors including the experience of the network operator in monitoring data traffic volumes.
The Network Operator can also influence whether other (lower-priority) objects in a change-log at the time of an open data connection should be replicated, once the connection-initiating object has been sent. It is possible to define an opportunism threshold to control this. E.g. several non-urgent items are in the device's change log. The user changes a device setting that immediately initiates a data connection. The NetOp has speάfied that any other object in the change-log under 20kB should be replicated using the open connection.
A different opportunism threshold should be used if the device is operating in a roamed network, as the cost to the user of initiating connections may outweigh the impact to the NetOp of sending more data at peak periods. The opportunism threshold, along with the other QoS control parameters on the client device, are communicated to the client using the replication framework and are therefore kept automatically in step between the client and the server.
In a more advanced implementation of QoS control, the network-based server will be able to determine the cell loading for a given mobile device. When the cell loading drops below a defined cell loading threshold, the server should be able to signal the client that it may start opportunistic replication. This delivers most benefit in all-IP networks where the client device has a permanently assigned IP address and can therefore be contacted instantly by the server. Otherwise the process of sending a Communications Initiation Request to the client device influences the cell loading, and adds a latency to the system that creates an opportunity for the cell loading to change before replication can start.
The Network operator can update the data traffic graph, opportunism threshold and cell loading threshold after deployment. This allows optimisation of the QoS control in the light of experience.
All objects joining the device change log are currently subject to the Pause Before Connect delay. However as a Network operator defined mechanism for defining to what extent changes are batched together, the Pause Before Connect delay will be largely superseded by QoS control. 2.2 The Service Provider control
The Service Provider has the opportunity to determine a QoS Profile for each Cognima service. This profile contains sets of replication time limits within which Cognima objects created by that service are intended to be replicated. The actual time at which replication is attempted within the time limits is determined by a number of elements in the system including the state of the Cognima client software and the client device, and network factors such as data tariff peak/off peak times, cell loading and so on. The Service Provider can set a zero time limit - i.e. request immediate replication for objects of a given type. If replication is attempted and fails, the object remains scheduled for replication but is subject to the existing backing-off behaviour.
E.g. Provider specifies that replication of new contacts to the Server should be attempted within a 2 hour time limit. If a contact is created during a network trough or while the cell loading is very low, the device may send the contact immediately. If the timing coincides with a network peak or the cell loading is high, the device can wait for up to 2 hours for conditions to change, but after this time it must attempt replication regardless.
The QoS Profile defines the following factors as parameters determining replication timing:
Object type (e.g. Contact, Photo, DCC directory, DCC usage log etc) How object was created (e.g. new contact on handset, edit to contact created on portal etc.)
Direction of travel (e.g. edits to contact on portal go immediately, edits to contacts on phone go within 2 hours)
Handset is on home/roamed network (e.g. photo replicates immediately at home, but within 12 hours if roaming)
And optionally...
Size of Object (e.g. contact is over 5kB (i.e. contains an image) — send within 12 hours, contact is under 5kB — send immediateyl)
■ Different behaviour when service is first activated (E.g. initial upload of photos is immediate, thereafter within 12 hours)
The time limit for replication is assigned to an object by a QoS Profile based on its time of creation. However it is also possible to change a limit already assigned according to subsequent events such as the memory available on device changing, or if the handset roams to another network. This creates a need for occasional recalculation of the weights of the items in the changelog.
The Service Provider is able to define a shelf life for objects in the QoS Profile. If an object reaches its shelf life while it is still in the change log, the object should be deleted.
E.g. The Service Provider specifies that a weather update in the DCC service has a shelf-life of 24 hours. If the item hasn't replicated to a handset within this time, it is deleted from the change-log
The Service Provider can define an object as over-writeable. If a new object enters the change log that replaces an earlier version still in the change-log, the earlier version is deleted. In this situation, the Service Provider can determine whether the new object should adopt the timing of the object that has been overwritten, or should enter the system with a new replication time limit. The default setting is that new objects adopt the timing characteristics of the objects they replace.
E.g. The Service Provider specifies that a objects of the weather update class are over- writeable, and that new entries should adopt the timing of the ones that get overwritten. This ' will mean that an old forecast message gets overwritten by a newer one, but the newer one is not held back from replication.
Service Providers can provide different classes of service within the Profile set. One way of dong this is by defining individual QoS Profiles for each class of service. E.g. Contacts in a Gold service are replicated immediateyl after creation, but under the Silver service on the same network, replication may take up to 2 hours.
A QoS tariff applies to an individual Cognima service, although it may be presented to the subscriber as a bundle of services at a particular price point.
A QoS Profile can be changed by the Service Provider once it is deployed; changes to the Profile are replicated to the relevant client devices so that the understanding of each QoS profile is common between all entities in the system. 2.3 Notes on deriving a QoS Profile.
The parameters in a network operator (or 'netop') defined profile should aim to balance:
What is acceptable user experience for given tariff? What scope should be given for batching changes (typically more batching while roaming)
What scope should be given for dodging network peaks
It should be expected that users will generally need changes made on a portal to be swifter than those made on the phone - we should expect users at a PC to have their phone with them, but not vice versa. Initial replication on activation should be immediate for the best initial user experience.
The cost to the user of opening several data connections when roaming is likely to outweigh the impact to the NetOp of sending more data at peak periods.
3. Algorithmic Implementation of QoS
3.1 Introduction
In the Cognima system, QoS is implemented as an increment to the functionality of both the client and the server. In particular it requires a modification to the behaviour of the change log, and introduces a requirement to recalculate certain attributes of queued change log entries. Replication can then occur as a result of the outcome of this recalculation.
3.2 Algorithm Description
The algorithm is made up of several components. We introduce a changelog item weight. This weight indicates how urgently a changelog item needs to be sent; the heavier the weight, the more urgent the item. We introduce a changelog threshold. Any items that have a weight equal to or greater than the threshold need to be sent immediately. When a connection is made all items with weight greater than the threshold minus a delta are sent. The delta represents opportunism.
For the sake of concreteness we say that weight and threshold can vary between 0 - 100. The weight of an item that absolutely has to be sent right now is 100; a changelog threshold of zero indicates that any entries in the changelog can be replicated right away.
Both weight and threshold can vary over the course of a day. There will be predictable variation and also dynamic variation. Some examples will clarify this.
In Figure 1, the straight line at weight value 40 shows that the weight is constant over time. The weight of an item that has to go within a certain time is shown at Figure 2. The weight of the item starts at a relatively low value, indicating low priority, and then it jumps as we reach the time limit - the new value of 100 will force the client to attempt replication of the object at time TI.
An item that should only go at a certain time will look like the Figure 2 diagram but the jump comes at a particular time, not after a particular duration. The weight of an item might change dynamically. Say for example that the device starts to run out of space: the weight of items could be increased to get them out of the changelog and allow ghosting (see Appendix one to the current description for an explanation of this term). The threshold will also have a graph over time, the following Figure 3 graph shows an example of how a daily threshold cycle might appear, with a high threshold to guard against low-value traffic during peak hours (e.g. after 9.00 hours for several hours), and a lower threshold when the data networks are traditionally quieter (between 24.00 hours and 6.00 hours, the threshold is zero).
The example effectively shows a daily cycle split into three tariff bands, perhaps named off-peak, standard and peak, with the replication threshold set appropriately for each band. Note that there is a fourth band just after midnight, where the threshold drops to zero - this is introduced to ensure that all changelogs are emptied once per day during the quietest period. This zero threshold period could be defined once per week or at some other interval defined by the network operator, but is recommended to ensure that the defined QoS profile does not prevent some objects from being replicated at all. As for the other bands in the profile, the off-peak period extends through the late evening and night, and represents the times at which the data network can expected to experience low traffic; replicating large objects during this time will allow' the mobile network operator to make best use of the limited bandwidth available during peak times.
The zero-threshold period can be adjusted by the network operator in the light of experience; finer adjustments can be made for different user groups, per-application or even per-user thereby staggering the off-peak network usage and ensuring that the full user base for a given server does not attempt to connect to the Cognima Server at the same time.
There will be dynamic changes to the threshold, for example from cell loading or to support a marketing drive promoting a new service (during which time it may be preferable to allow basic tariff data onto the network during peak hours to encourage uptake). If it is possible for the device to detect that its cell is not busy it could drop its threshold a bit which might trigger some replication.
The core of the algorithm is to calculate the threshold graph and the weight graph of every item in the change log. If the current weight of any item is greater than the current threshold then a connection is established. Otherwise the next time that any item will exceed the threshold is deduced (ie when the graphs intersect) and a timer is set for this interval. Because both the weights and the threshold can be dynamic there are several events that can trigger a recalculation:
• A new change log item is added
• The server may push a new threshold value to the client. This is actually a special case of the previous event since the QoS object on the client is controlled through replication in the normal way, meaning that a new threshold value will be delivered to the client by placing the change in the device's changelog queue on the server - this changelog entry must have a weight of 100 to force immediate replication, and the resulting change of threshold may trigger or delay replication of other entries already in the queue.
• A timer expires — this will usually be the timer indicating the point at which the weight of an existing entry in the changelog is due to exceed the current changelog threshold. • The cell (or network) loading changes
• The memory available on the client device falls below a certain level
• The device detects that its roaming state changes
• A new application is deployed and activated on the device • Connection terminated — this also results in the creation / update of the 'next connection' timer. There are two separate calculations: the weight graph of an item and the threshold graph.
Parameters that can affect the weight of an item at any given point in time are: • Direction (client -> server or server -> client)
• Shelf life (this is usually encoded in the class)
• Overwritable (this is usually encoded in the class)
• Size in bytes
• Time entered into change log • Priority
• Time out interval
• Assigned time for replication
• User assignment of a non-default priority to a given object (such as the 'send now' option on an image for uploading to the user's media gallery account) • Memory available
The extent to which these parameters influence the weight is controlled by the service provider (i.e. network operator) in the control as described in section "Service Provider Control" above.
Parameters that can affect the current threshold value of a change log are:
• Time of day
• Roaming status
• Cell / network loading
• Time since last replication • User tariff
After each refresh of the changelog, the client software also calculates the time interval to the next intersection between a weight graph and the threshold graph; this is so that scheduled replication can take place as required even if there are no further changes before that scheduled time. Generally, the 'next event' for the change log will either be as predicted by this intersection calculation, or will be created by some external event that places a new entry in the change log (which may of course force a refresh of the values of all weights and the threshold). Note that the 'next connection' event may have a value of 'never' if the changelog is empty, unless the active QoS Profile has a zero-threshold period as in the example above.
3.3 Banding
We control both the client weight graph and the threshold graph by a structure we call banding. We assume that all graphs can be described as sets of horizontal lines which jump (concatenations of step functions). Therefore the QoS Profile for a given class can be described by an array of bands. A band is parameterised as follows:
Figure imgf000016_0001
Bands can be defined as deltas from a start time or against the system clock (absolute). A weight graph for a class which should be scheduled to go within 2 hours of object creation could be described by a pair of bands: The following table could describe the QoS weight profile for the Contacts class.
Figure imgf000016_0002
A threshold graph which describes the graph in the example above would be represented as follows:
Figure imgf000017_0001
Note that there is a different weight for objects created or modified on the server and that this weight is constant throughout the daily cycle. This approach can be used to mirror the fact that server-side changes may have a different priority to those originating on the client.
The client and server carry QoS objects which encapsulate the above tables and influence the scheduling of replication. A user's device will hold a QoS object for each data class installed on the device, plus a single threshold object representing the whole system. Each QoS object holds an array of bands to describe a single weight graph. The intersection points of these graphs determine when objects of a given type will be replicated - these intersection points must be calculated by determining the weight of each object and the threshold value for the system at a given time. 3.4 Changelog item weight calculation
The weight of an item in the change log has to be calculated. The QoS object for the item's class is retrieved from the data store. The banding structure is examined and a weight is looked up from the weight graph (either by calculating current time — created time in the case of delta band type, or by comparing system time in the case of absolute band type). To include object size as a parameter in weight calculation, we have added two more fields to the banding structure.
Figure imgf000018_0001
If the object size is greater than the size limit then the oversize weight is used — note that this could be higher or lower than the default weight for the class. In some applications, the object weight can be overridden by a user request (effectively a 'send now' button) which sets the weight to 100.
3.5 Changelog threshold calculation
The current changelog threshold weight can be extracted from the threshold QoS object using the current system time. This value can then be modified by dynamic variables — for example if the device can detect roaming status then this can influence the threshold. Generally the threshold will be higher when roaming, to reflect the fact that replication will be more expensive; it is also possible to specify a lower limit for the threshold in a given band, effectively preventing very low priority objects from replicating at all before the device returns to its home network. In deployments where the mobile terminal can be aware of local cell loading conditions, then the cell loading can be used as a factor in adjusting the current threshold value: if the traffic loading of the local cell is below some value (e.g. 70%) then the threshold can be reduced. If the cell loading is above some value (e.g. 95%) then the threshold can be increased.
The QoS object creates timer events which represent the times of day at which the threshold is known to change, reflecting the shape of the QoS Profile as defined by the network operator. As each of these timer events fires off, the QoS threshold will be assigned a new value and the ensuing weighting recalculation will allow objects with the correct weight to be replicated.
3.6 Next connection time calculation The weight graph of an item needs to be compared to the change log threshold graph to find the next time that the item weight >= threshold weight. This calculation will ignore dynamic change to the threshold weight which is by definition unpredictable. It is a fairly simple calculation to match up the bands and compare weights at consistent times.
There are three types of event that might result in a replication session: • a change object already in the changelog may increase in weight due to moving into a new band
• the QoS threshold may drop due to a move from one time-based tariff to another
• a new entry may appear in the changelog with a weight higher than the current threshold.
For the first two events, the time at which the session starts is predictable and must be calculated at the end of each replication session. This represents the next scheduled connection time.
APPENDIX 1 DATA REPLICATION SYSTEM DESCRIPTION
The present invention will be described with reference to an implementation from Cognima Limited of London, United Kingdom. Cognima has developed a data replication technology that directly addresses the need for Mobile Service
Providers (MSPs) and Network Operators to increase consumer adoption of data services, encourage greater loyalty from their valuable customers, and differentiate their services from the competition.
Cognima's data replication solution addresses these issues by: • Increasing adoption by making data services compelling and effortless to use.
• Establishing a high barrier to churn by securely backing up subscribers' personal data on servers controlled by those subscribers' MSP.
• Enabling the MSP to create differentiated services by controlling the customer experience.
1. Overview of uses for the Cognirna Data Replication Framework
Cognima's data replication framework enables a Mobile Service Provider to build compelling services for consumer markets. The MSP hosts a Cognima Server at its data centre. The server comprises an Oracle database plus
Cognima's multi-threaded Java communications server, hosted on a standards- based J2EE application server and carrier-grade Unix hardware. Section 4 and later sections describe the technical implementation in detail.
The Cognima framework replicates data entered in a mobile phone automatically
(without any user intervention) to other phones via the Cognima Server. Similarly, data from external systems connected to the Cognima Server is automatically kept up-to-date on mobile phones.
Mobile subscribers using Cognima-enabled applications experience an always- available, instant connection to their personal information and friends. Personal information can include the subscriber's address book, messages, bank account details, stock prices, pizza orders, calendar, current traffic on a route to work, or any other personalised content. The data is always kept securely backed-up on the Cognima Server and automatically replicated on all relevant client devices.
• Always-available means that the personal information is accessible on whichever device or handset the subscriber is carrying, whether currently connected to the network or not since the user can always access personal information stored locally on the device). Users can also edit and manage their personal data directly on the server via a web interface - the Virtual Phone.
• Instant means that subscribers do not have to wait for data to download from a server; the latest information is on their handsets even before they know they need it since that data is automatically sent to the handset (e.g. polling by the handset may occur; this can be regular periodic - such as every 30 minutes or at pre-defined times (4pm, 5pm etc). Pushing to the handset may also occur).
• Subscribers can share their data across multiple devices and with their friends since the Cognima Server can replicate this data to any defined device or defined individual.
1.1 Example Cognima Applications
Figure imgf000021_0001
Figure imgf000022_0001
Figure imgf000023_0001
Figure imgf000024_0001
Benefits to the Mobile Subscriber
Cognima provides an ideal framework for implementing mass-market consumer data services based on the following key benefits:
• Friendliness: no user intervention is required. Subscribers never need to press a 'sync' or 'download' button to access their data. System configuration and secure data transfer are completely transparent to the end user.
• Instant availability: the user is always able to interact instantly with local data (even when off-line), whilst any updates take place silently in the background. For example, users can read their personalised content whilst on an underground train. The user experience is separated from the data transport.
• Affordability: The MSP can control when replication takes place, and the Quality of Service (QoS) delivered. However, because the user experience is separated from the data transport, lower QoS does not affect the user's perception of the service. Crucially, this allows the MSP to offer low-cost, subscription-based services with relatively poor QoS without sacrificing user experience - e.g. data replication can happen overnight for non-urgent data services such as bank statements, yet still be satisfactory to users. Overnight data replication uses otherwise underused bandwidth and is hence far cheaper than peak time data replication. Urgent data replication (e.g. presence information) can happen at any time on a periodic or (optionally) continuous (push) basis and attract a higher charging rate. Furthermore, efficient use of phone memory & processor power allows Cognima client software to be cost-effectively installed in even the cheapest mass-market phones. 3. Benefits to the Mobile Service Provider
Cognima presents a MSP with a means to generate new data revenues, reduce churn, and to differentiate its services from those of its competitors.
3.1 Increased Usage of Existing Mobile Services
Cognima increases usage of existing mobile services:
• Messaging and content-based services become much more convenient and immediate, and will therefore be used more. • The enhanced immediacy of presence information increases the use of chat and Instant Messaging, and an alert when free capability will boost voice calls.
• Effortless management of multiple devices allows users to carry an appropriate phone on any occasion, and therefore make more calls and send more messages.
3.2 Compelling New Services
Cognima enables rapid introduction of compelling and affordable new mobile data services. • Cognima delivers a compelling user experience for new services in low-end phones using only spare network capacity. This is affordable and scalable for the network operator, allowing the MSP to offer understandable and predictable pricing for mass-market subscribers.
• Most of the application development for new Cognima services takes place on the server side, allowing the MSP to bring new services to market quickly.
• Cognima's client software can be installed as a flash memory upgrade, endowing today's mass-market handsets with smart-phone-like capabilities. New software applications can be downloaded over the air to existing Cognima-enabled handsets, allowing MSPs to roll out new data services without waiting for new devices to support them.
• Third party application developers can leverage the MSP's Cognima infrastructure to develop new applications for the MSP's network. 3.3 Churn Reduction
Cognima services act as a significant barrier to churn. For example, a subscriber who stores their personal information securely at their MSP's Cognima Server can buy a new phone and immediately retrieve all personal information to their new device. All this personal information may be lost if they decide to take out a subscription with a different service provider.
3.4 Differentiation
Today, subscribers have the same basic experience of using mobile data services on all networks. For example, the experience of using WAP services is defined by the WAP protocols, the browser in the phone, and the content accessed. Many MSPs have realised that they must differentiate themselves by giving their subscribers a unique user experience, but are hindered from doing so by severe constraints to customising the services in mobile handsets.
Cognima gives MSPs the ability to implement services on the handset, and thereby to regain control of their subscribers' user experience. Most importantly, Cognima allows this without sacrificing interoperability; support for industry standards is achieved through straightforward integration with the Cognima Server. The net result is that the MSP's position in the value chain is strengthened versus the powerful brands of handset manufacturers and content providers.
4. Cognima data replication framework functional design
4.1 Introduction
This and subsequent sections of the Detailed Description are intended to describe how the Cognima data replication system actually works. It covers the behaviour of client devices, the Cognima Server and the web client, without going into details of specific hardware, programming language, software class design or environment. It does describe the basic data structures and algorithms used. Terms
Figure imgf000027_0001
4.2 Purpose
The objectives of the Cognima software are:
• To allow a user instant access to view and modify an 'up to date' copy of their data on multiple handheld devices capable of wireless data connectivity.
• To allow a user to view and modify the same data using a conventional web browser.
• To effortlessly provide secure backup of a user's data.
• To give a user powerful data functionality on a cheap handset by displacing complicated and expensive processing to a server.
4.3 Highest level description
Client devices hold a copy of the user's data in a database on the client device. The user can access this data whether or not he has a network connection and therefore always has instant access. When a user changes the data on his device, the changes are copied to a Change-Log. The client device connects periodically to a Cognima Server on the wireless network, to send up the changes from the Change-Log and receive new data. This separates the act of changing data from the need to connect to the network (i.e. push is not continuous in a preferred implementation). The Cognima Server updates its own database with data changes received from the client device, and populates Change-Logs for any other devices the user owns. When these devices next connect, they will receive the changes and thus the devices are kept in sync, each with a copy of the same data.
The Cognima Server contains a web server which allows the user to examine directly using a web browser the copy of the data held in the Cognima Server database, and make changes to it as he would on a client device. The Cognima
Server also acts as a gateway for the user to communicate with other servers on the network / internet. For example, the client device can effectively ask the Cognima Server to send a message as an SMS or an email or a fax by setting a few flags in a message object and the Cognima Server contains the functionality to communicate with email servers, SMS servers and fax machines. This can be extended to servers holding ringtones, banking details, games etc. It is easier and cheaper to build the software on the Cognima Server to talk to these other servers, than it would be to build the software on the client device.
5. Lower level concepts
5.1 Data structures
5.1.1 Ids
Cognima user data is described using the terminology of object databases: classes and objects. Unfortunately, there is room for confusion with similarly named OO programming concepts and care therefore needs to be taken.
All users in a Cognima network are assigned a user id. This id is unique to the network — i.e. provided by a given network operator. All users have a Cognima address which is a combination of their user id and Cognima Server URL. This is unique in the world. Each device which belongs to a user is assigned a device id. The device id is unique to the user. This is only 8 bits so a user can have a maximum of 253 devices (id 254 is reserved for the web, id 255 is spare, id 0 is invalid). All user data is classified into classes (contacts class, messages class, bank transactions class etc) and the classes are assigned a class id which is unique in the world. Class id '12' refers to a contact, for example.
An instance of a class is an object, which is assigned an object id unique to the user, e.g. a contacts class object might be the contact for "John Smith". The object id is generated by concatenating the device id of the device which created the object with a monotonic increasing count which increases over the life of the device. So each device can create a maximum of 16777215 objects (if we encountered this limit we could reset the device id). Classes are defined by the properties which constitute them. A class is essentially an array of properties. Each property in the class has a property id which is unique to the class (and is actually just the array position of the property in the property array, starting from zero). 5.1.2 Creating objects
An object is created on a device. It is assigned an object id and saved to the device database. A copy is also saved into a Change-Log. When the device next connects to the Cognima Server the entry in the Change-Log is sent up. The Cognima Server saves the object to its database (recording the system time), does any class specific processing that may be required (such as generating and sending an email) and adds entries to Change-Logs for any other devices that the user may own which have declared interest in the class. (The entries should be for the correct version of the class on the device).
An object may also be created on the web portal. The object id is generated
(using device id of 254 as described above) and processed identically to the device. There is no Change-Log for the web portal, it gets selections directly from the Cognima Server database.
An object may also be created by a server application (e.g. a messaging module might receive an email from which it creates a message object). The object id is generated (using device id of 254 as described above) and processed identically to the device.
5.1.3 Updating objects
One or more properties of an existing object are modified on a device. The changes are saved to the device database. Each changed property is used to generate an entry in the device Change-Log. These are sent up to the Cognima Server.
If the time of the update is later than the 'last changed' time for the property in the Cognima Server database then the Cognima Server saves the changes to its database (recording the new ast changed' time for the property), does any required class specific processing and adds entries to Change-Logs for other devices which belong to the user, have declared the class and have a version of the class which contains the property. The update is also placed on the Change- Log for the device that originated the change. This may seem strange but is required to cope with the following scenario:
A user has 2 devices A and B. He updates property 7 on A offline at 5pm and updates it on B offline at 6pm. He connects to the network with A first. The value of 7 on A gets put in the Change-Log to be sent to B. Later B connects. Its value of 7 is more recent so the value of 7 on B is sent to A, but B getsA's value. Replicating the value of 7 back to B fixes this.
If an update is received by the Cognima Server for an object which is marked as deleted and the update is later than the deletion, then this is interpreted as an undeletion. The object is undeleted, updated and then a refresh of the object in placed on the Change-Logs for all appropriate devices. Updates from the web portal or server applications work in the same way.
5.1.4 Deleting objects
An object is deleted on the device. It is removed from the device database and an entry is put on the Change-Log listing the class id and object id. The entry is sent up to the Cognima Server. If the time of the deletion is later than the last updated time of the object, then the Cognima Server marks the object as deleted in its database, does any class specific processing and adds the entry to other devices that belong to the user and have declared the class.
If the time of deletion is earlier than the last updated time then this indicates that the deletion is invalid and a refresh of the object is put on the Change-Log for the device which originated the deletion.
The deleted object is viewable in the web portal a manner that makes its deleted status clear. The user can select the object for un-deletion. The deletion mark is removed from the object in the Cognima Server database and entries to refresh the object are placed on the Change-Logs for all devices that belong to the user and have declared the class.
5.1.5 Property types
Each property has a type. There are currently 9 permitted property types:
Figure imgf000031_0001
Figure imgf000032_0001
A CogString is a character count followed by the characters. If the string is ASCII then the space taken up by the string will be (4 + char count) bytes. If the string is Unicode then the space taken up will be (4 + (char count * 2)) bytes. A CogTypedString is a CogString preceded by a type (4 byte unsigned integer). The only use of a typed string so far is a Contact Point. The type identifies the type of contact point (e.g. email address, home phone) and the string holds the address (e.g. bob@xxx.yyy, 01233556677).
A CogBlob is a length in bytes followed by that number of bytes. It can be used to store any binary data.
A CogArray is passed around as a 4 byte unsigned integer 'type' followed by two blobs. The 'type' indicates the type of elements held in the array. The first blob is an index blob: it holds a sequence of offsets (4 byte unsigned integers) into the second blob. The second blob is the data blob which holds the elements of the array as a sequence of binary lumps. Elements can be extracted from the data blob by counting along the index blob to get the offset of the start of the element in the data blob. This is the stream structure of the CogArray as it is passed around. Inside a particular system it may appear as a conventional vector (i.e. already parsed).
The only implemented example of a CogArray is the MessageAddress. Each element of the MessageAddress is an AddressPair. An AddressPair is a contact id (object id of a contact object) followed by a Contact Point.
5.1.6 Smart property parameters
Some of the properties can be made "smart". This means they can be parameterised for a specific device to sculpt the data in the property for the characteristics of the device. In practice the parameters are two 4 byte unsigned integers, one is a smart type and the other is a max size. For example, the property which holds the body text of a message might be parameterised to smart type kCogPlainText and max size 100 on a cheap phone with limited memory, but parameterised to be smart type kCogRichText and max size 1000 on a PDA with more memory.
The parameters are stored by the Cognima Server when the application is added to the device. When new objects or updates for that class are placed in the Cognima Server Change-Log for that device they are processed according to the, smart parameters. This might involve, for example, truncating text, converting Unicode text to narrow text or converting image formats.
It is important for data integrity that the object held in the Cognima Server database be a copy of the object as it was generated. Even if you see a cut down version on a device you can effectively manipulate the complete version on the Cognima Server.
5.1.7 Class versions We have the concept of a class version which is defined by a 4 byte unsigned integer. A new class version may add properties to the end of the old class, but it may not change or remove existing properties, or insert new properties between existing properties. This should allow interoperability between versions. Class definitions with different smart property parameters are not different versions.
5.2 Passing user data around
Cognima utilises the idea of class metadata to minimise the data that needs to be copied around between databases. Class metadata is essentially an array of property metadata. Property metadata is a property id, a property type, a smart type and a max size. User data is transferred as a stream with no formatting information other than a class id. This stream is parsed by looking up the class metadata. So if a stream is received for class 6 and the class metadata for class 6 says that property 0 is a KcogTypeUInt and property 1 is a KcogTypeStr, then you know that the first 4 bytes of the stream should be interpreted as an unsigned integer, the next 4 bytes should be interpreted as an unsigned integer holding the number of characters n in the succeeding string, the next n (times 2 if Unicode) bytes hold the characters in the string etc.
Client devices declare to the Cognima Server the classes that they support. This enables the device to subsequently send up only raw user data (with a header containing class id, object id and a few other things) and hence minimises bandwidth requirements. This can be contrasted with, for example, XML reliant systems that are far more bandwidth hungry.
The client device class declarations also contain the smart property parameters so that the Cognima Server can sculpt the data for the device. It is worth emphasising that the meaning of a property is hard coded into an application. The class metadata states that property 2 in class 7 is a string with max length 30 characters. It is the code in the application that interprets property 2 in class 7 as the name of a football team.
5.2.1 Data replication issues in more depth
Data is held in objects that are created on client devices and the server these devices connect to (known as the Cognima Server). These objects and any changes made to them are replicated between the client devices and the Cognima Server. The design of the replication process allows:
• A set of objects to be defined that will be replicated so that the same set of objects will be held on a Cognima Server and all the client devices that are logged on to that server for a given user. New objects created on any device or the server will be replicated to all other devices.
Changes in any property of an object will be replicated to all devices. • Only the minimum data to be transmitted across the network for a given update since only changes in data are sent from clients to the Cognima Server or vice versa.
• A key part of the design was to not require times of modification to be kept for each property of an object on the client device as updating these on constrained client devices is slow and keeping a last modified time for each property in an object would take a lot of space.
• On the Cognima Server storing modification times for all properties of an object is fine as the server has enough storage space and processing power to deal with this.
5.2.2 Metadata
In order for the system to work it needs a clear idea of what properties are defined for a given class of objects. This is done by providing the programmer with a few C++ compiler macros that allow definition of the class metadata.
The definition of the properties to be used in a class result in a Class Metadata definition. This definition tells the CRE (Cognima recognition engine) what type a given property is and allows it to pack and unpack an object or a property for transmission over a data link. In order for the CRE system to work all clients and the server must have the same class metadata definition. Thus the following occurs:
• When a new Metadata definition is declared on a client device it is sent to the Cognima Server and from there the Cognima Server will send it to all other clients. • When a new Metadata definition is declared on a Cognima Server the definition is sent to all client devices.
• When a new client device logs on to a Cognima Server for the first time all of the metadata definitions are sent to that device before any objects are sent. • In all of the above cases a future optimisation may be made so that the
Cognima Server only sends the metadata definition to clients who access the class (and the specific properties) the metadata refers to. 5.2.3 ChangeLog
The purpose of the ChangeLog is to record any changes that have occurred since the client device last connected to the Cognima Server (or the Cognima Server to the client device). Using Cognima APIs, applications connect to the CRE and can cause objects to be created or deleted, or a property in an object to be changed. These changes are added to a Change-Log on the local device as they are made together with the time the change was made. Objects are given unique identifiers when they are created so that a given object can always be identified.
In the same way, creation and deletion of objects and changes to object properties by applications running on the Cognima Server result in the changes being added to all the Change-Logs of all the client devices registered to that user on the Cognima Server. The time of changes are recorded for each object or property.
ChangeLogs can be built in two ways:
• As the new objects are created and properties are changed (this is normally the case for client devices)
• Or they can be built on demand when they are needed by using the last modified times of objects and properties if these are stored on the system (in some circumstances, this method may be used on the Cognima Server instead of the above method).
5.2.4 Replication
When a client device has items in its ChangeLog to send it will connect to the Cognima Server (and likewise for the Cognima Server connecting to the client device). By default, the items in the ChangeLog are sent in the order in which they were added to the ChangeLog, however they may be re-prioritised immediately before sending to provide for premium services, urgent data and so on. Items transferred are: • A metadata definition including the type of each property of a given class of objects. A new object that has been created - with the contents of the properties of that object.
• A property has been changed - with the new value of the property.
• An object has been deleted.
In all the above cases the appropriate IDs are sent to identify the object, class and properties involved. All ChangeLog items are marked with the time the item was added to the ChangeLog. These times are always local machine times and are resolved into GMT by the Time Management approach described in Section 6.2.
When a client device receives ChangeLog items from a Cognima Server:
• When a client device receives a new object message from a Cognima Server it adds this new object to its local database.
• When a client device receives an object deletion message from a Cognima Server it marks the object as deleted in its local database.
• When a client device receives a property change it is always assumed that the Cognima Server is authoritative on the current state of the database and so the change is always made to the value of the property held in the local database.
A Cognima Server receives ChangeLog items from a client device:
• When a Cognima Server receives a new object from a client device it is added to the Cognima Server database and also added to all the Change-Logs of the client devices registered to that user, apart from the Change-Log of the machine that sent the new object in the first place.
• When a Cognima Server receives an object deletion from a client device the object is marked for deletion and an object deletion message is added to all the Change-Logs of the devices registered to that user apart from the Change-Log of the machine that sent the object deletion in the first place.
• When a Cognima Server receives a property change it compares the time of the change to the current time held for that property on the Cognima Server. If the time of the property change is later than that held on the Cognima Server the property value is changed in the server database and this change is also added all the Change-Logs of the client devices registered to that user - including the one of the machine that sent in property change (in case another object update has been sent to that machine in the meantime). If the property change was not later than the one held on the Cognima Server no change is made as the stored property value is more recent - but the value is added to the list of old property values on the Cognima Server so that a user can retrieve it later if required. When times are compared the Time Management approach described in Section 6.2.below is used.
When a device first connects to a Cognima Server it will be sent all class metadata definitions and then all the objects in the database for that user. The Deletion messages generally just mark an object for deletion. Actual removal of the object from the database may occur later on once all objects referring to that object have also been deleted.
5.2.5 Optimisations
An optimised version of the above replication protocol allows for aggregation of the entries in the ChangeLog. If a ChangeLog (in the Cognima Server or on a client device) has not yet been replicated, and a subsequent entry is added, then existing entries can be scanned to potentially reduce the number of entries that need to be replicated during the next connection:
• if the new entry is an update to a property that is already scheduled for update then only the later entry need be retained • if the new entry is an object deletion then all property updates for that object can be removed from the ChangeLog
• if the new entry is an 'undelete' command and the original deletion is still in the ChangeLog then the two entries can both be removed from the ChangeLog 6. Core algorithms
6.1 Handling endian-ness
Operating systems are fundamentally little endian or big endian which is a choice of the byte order in which numbers and strings are stored. If two computers which have different endian-ness have to communicate then one of the computers will have to switch the endian-ness of its data packets. In the Cognima environment the Cognima client software uses the same endian-ness as the host client device. The Cognima Server has to determine the endian-ness of the client device (it uses a reference value in the first packet of data from the client) and then convert the subsequent incoming data if necessary to maintain consistent endian-ness in the Cognima Server. The Cognima Server also has to convert any outgoing data it sends back to the client device.
6.2 Synchronising system times
Different devices will inevitably have slightly different system times. Changes that are sent from client devices to the Cognima Server are stamped with the device system time at the time of the change. It is up to the Cognima Server to resolve the times on different devices so that it can judge the order in which changes took place and record the correct update.
The logon of a device contains the current device time. The Cognima Server should be able to compensate for the latency of the network and compare the login time with its own system time. This will give it a delta between the device time and the Cognima Server time. This delta can be applied to further times sent up by the device in that session.
The Cognima Server can compare deltas in successive sessions from a device to determine clock 'creep' on the device or changes of time zone: it cannot be assumed that all the client devices in the system have clocks that are well synchronised to each other: • Clock times drift on devices depending on the device's clock accuracy.
• Some users like to set clocks 5 minutes early for example. Some users will make changes to clocks to account for daylight saving rather than adjusting the locale settings (and some OSes may not provide locale features anyway forcing the user to change the clock direcdy).
To get round this problem, the server will be responsible for adjusting times used by the client device to GMT when comparisons are made on the Server, and from GMT to the equivalent time for the client device when messages are sent from the Cognima Server to the client device.
The client device will tag all the items in the ChangeLog with times obtained from the local clock - as far as the client device is concerned it only ever deals in time based on the client device's own clock.
Each time the client device connects to the Cognima Server it sends its view of the current time as given by the clock on the client device. From this the Server can work out:
• What the delta to GMT is
• If there has been any drift in the mobile device clock since the last time it logged on since the server keeps a record of the last delta to GMT and when the last connection was made and therefore can compare these. If there is drift the server can adjust all times sent by the mobile device pro-rata.
For example the table below shows a pattern of events with a client device connecting to a Cognima Server. The Client device's time is 5 minutes slower that the Cognima Server and is loosing a minute every hour (an extreme case to show the point). Also to show the point we will assume that from 09:00 to 12:00 the user is on a plane and out of contact with the Cognima Server so it does not connect during this time:
Figure imgf000040_0001
Figure imgf000041_0001
In order to work out if the property changes were made before or after the time stored on the Cognima Server the times X and Y need to be worked out. From the information above the Cognima Server knows that when the client last connected it was around 3 hours ago and at that point the time difference was 5 minutes whereas now it is 8 minutes. Thus, assuming the clock drift happens linearly, the Cognima Server can work out that the device is 5 minutes behind GMT and that the clock is drifting back a minute every hour.
From this is it possible to work out that the time the client device knows as 10:00 for the property A change needs to have 5 minutes added to it for the initial drift, plus one minute for the extra drift that occurred in the hour till that property was changed.
Likewise Property B needs to be adjusted to 11:07 - the 5 minutes initial drift plus 2 minutes since two hours elapsed from 09:00 to 11:00 when the property was changed.
In practice the delta to the time between the client device time and GMT may be minutes, but the drift will be in the order of fractions of seconds per hour.
6.2.1 Time Adjustments
As well as the delta to GMT and any drift in the client device clock, users can also change the time on the client device. They may do this to reset the time to the correct local time (we can give the user the option to have this happen automatically but some users may want to keep their own control of their client device time — e.g. they like to have the clock set 5 minutes fast). They may also make adjustments to reflect a change of local time (i.e. daylight savings or changing timezone). The goal is that the user can change the clock on the device to any time that suits the user and the device simply takes account of this.
When the user makes a change to the client device time most operating systems will report this change (for systems that don't do this the time can be polled say every minute to check for such a change). On detecting a change in time the client device will work out the delta between the new time and the time as it was before the change. For example this may be a change of plus one hour as a user moves timezone. The client device stores this time difference as the Adjust Time which it saves for the next connection to the Cognima Server. The client device also goes through every entry in the ChangeLog and updates all times in the log by Adjust Time. This ensures that the entries in the ChangeLog are always relative to the local time on the client device.
Several such adjustments could be made between connections to the Cognima
Server — each time the amount of the time change is summed with the Adjust Time and the ChangeLog updated so that the times in the log are all relative to the local time on the client device.
When the client device next connects to the Cognima Server the client device sends at logon the stored Adjust Time — i.e. the amount by which the client device clock has been adjusted backwards or forwards since the last connection. The Cognima Server can then remove this amount from the time from the delta to GMT and drift calculation.
6.2.2 GMT to Client Device
The same set of calculations can be made in reverse to convert the GMT times of changes made on the Cognima Server to the correct local time for a given client device. 6.3 Adding an application
An application will use one or more classes to hold user data. The definition of the class is hard coded into the application. The version of the class is coordinated by releases of the application.
Say that a statistics application uses a Footballer class to hold data about footballers. When the application starts on a client device for the first time, it inquires from the device what version of the Footballer class the device already holds. If the version on the device is the same as the version that the application has been hard coded to use then nothing more need be done.
If the device holds a newer version of the Footballer class, then the application needs to be robust enough to cope with more properties than it expected. (This situation would arise if you had a class being used by multiple apps and for some reason you installed an older version of one of the apps. This should be rare: ideally interdependent apps should be upgraded together.)
If the device holds an older version of the Footballer class (or no version at all) then the application's version of the Footballer class should replace it. The new version is sent up to the Cognima Server. The Cognima Server therefore maintains a list of versions of classes used on all devices.
The web portal pages will be the equivalent of the hard-coded device application. The web can extract objects from the database according to the latest version of the class, and if there are more properties than it was hard coded to expect it can ignore them. Therefore the web does not need to declare class versions.
6.4 Change-Log optimisation
The Cognima Server maintains Change-Logs for all devices listing changes that will be sent to the devices when the devices next connect. There will be optimisations that can be made to the Change-Logs, for example: If >2 updates to the same property are queued in the Change-Log then only the last need be kept.
• If a deletion is queued for an object then any updates ahead in the queue may be removed. • If an update is queued for an object then any delete ahead in the queue should be removed.
• If a device registers a new application there could potentially be very many objects to send down to it (e.g. message history). The Change- Log should only have a sensible number of objects added to it (e.g. the 20 most recent messages).
7. Ghosting, resurrection, pinning and withdrawal
The space available on a client device to hold user data will typically be orders of magnitude less than the space available on the server. The device needs to hold a subset of data and the user should have to do as little work as possible to maintain this subset. Ghosting and withdrawal are tools to aid this.
A class definition may include flagging certain properties as 'ghostable'. This means that if the object is ghosted those properties will be nulled, freeing room on the client device. Ghosting is done automatically on the device. The decision about' which objects to ghost is made by following a 'ghosting rule' and applying the rule whenever an object is created or updated. The rule defines the maximum number of a selection of objects. When the maximum is exceeded the objects in the selection at the bottom of a sort order are ghosted.
For example, the class might be messages, the selection might be messages in the inbox, the sort order might be by date/time and the maximum number might be 50. If there are 50 messages in the inbox and a new message arrives, the oldest message in the inbox is ghosted.. Ghosting may remove the message body but leave enough- header information for the message to be recognised. Withdrawal (also known in the past as auto-deletion and removal) is similar to ghosting but works by removing the entire object, not just part of it.
Neither ghosting nor withdrawal are notified to the Cognima Server. They are purely local to the client device. Therefore different devices may have different numbers of objects. The data on the devices is still fundamentally in sync, but the devices hold different data subsets.
If the user wants to resurrect a ghost then a request is passed from the client to the Cognima Server for the object to be resurrected. A refresh of the object is sent down to the device and the object is put back to normal.
Individual objects can be pinned. A pinned object is never ghosted or removed. Pinning can be chosen by the user, or it can happen automatically. For example, an object that is resurrected is automatically pinned.
8. User replication - sharing objects
There are many applications for which we envisage it will be useful for users to be able to share objects. The general way that this will work is: A user needs to know the Cognima address of users that he may want to share objects with. It is more appropriate to discuss the retrieval of these addresses in detail in the Cognima Server architecture. Here we assume that such a list is available.
A set of one or more Cognima addresses is attached to the object which is to be shared. The object can be set to read-only (so the people you share it with cannot modify it). When the Cognima Server receives the new object (or receives an update to it) from the web or a client device it replicates it as normal.
It also looks up the list of 'sharees' Cognima addresses. It marks the object with an originator id (i.e. the Cognima address of the object owner + the object id) and sends it to the sharees. The sharee users may exist on the same Cognima
Server or be on different Cognima Servers. The Cognima Server of the sharee receives the object. If it is a new object it assigns a new object id (keeping note of the originator id). If it is an update it finds the object using the originator id. If the sharee is allowed to update the object, the update can be replicated back to the object owner using the originator id.
9. Displaying data
Conventional small devices like PDA tend to have simple filing systems that allow applications to read and write data to some kind of storage that will keep the data when the application is not running. Generally these programs will tend to read in the available set of data and then provide a user interface to display the data on the screen. This has some disadvantages:
• Reading in the data when the program starts takes time
• The application needs to store all or some of the data in memory meaning it is now occupying more memory on the client device • Allowing more than one application to access the same set of data becomes non-trivial
• Similar code to read and manipulate the data appears in several applications that run on the device.
The Cognima approach is different: • Data is stored in an Object Database that can be accessed by several applications
• A Cognima application does not read in all the data it deals with from a database. Instead it creates a selection — a subset of the data which it is currendy interested in. In general this selection matches the data that is currendy being displayed on the devices screen. Thus only the data currently being used by the application is held in memory - saving a lot of memory space.
• All of the work of storing, sorting and indexing the data is done by the Object Database and so this functionality does not need to be repeated in each application.
• When changes need to be made to data in an application, the application never direcdy updates its own display of the data. Changes will update the properties in an object or create or delete an object. A change to the data could be made by another application or an update received from a Cognima Server due to the data being changed on another machine. • When an application sets up a selection it gives a list of criteria by which data is either included or excluded from the selection - because of this the Cognima Replication Engine can tell which applications to notify when a object is created, deleted or updated.
• When an update needs to be sent to the application, code in the application linked to the selection that contains this data is called and in this way the application can respond to the changes that have been made.
• When selections are set up, the application can also specify how the data is sorted and if only a small window on the sorted list of data is required (known as a view).
This approach is similar to the screen re-paint approach used to redraw graphics screens on Windowing systems. When an area of the screen needs repainting the application that is responsible for that bit of screen is called to repaint the screen.
9.1 Example A client device may have a contacts application running on it - this device replicates data with a Cognima Server connected to other client devices also running contacts applications. A class of object is defined for a Contact that contains names and phone numbers and these are replicated to all the devices of a given user.
An application on one device may have a display that shows all contacts by beginning letter - for example the interface allows the user to press a D button to show all the names beginning with D. This application will set up a selection that contains objects: • Where the class is defined as Contacts
• Where the name begins with the selected letter (e.g. D) When the selection is defined the application also defines code to be called by the CRE when objects are added, deleted or updated.
When the selection is first set up this code will be called back with the first set of objects that fulfil the above criteria.
If the application was asked to create a new contact with a name beginning with D the application would create the object but do nothing else. The CRE would detect the new object and call back the selection code to notify it of the new object.
Likewise is a new Contact object was created on another device and was replicated to the client device - if the name of that Contact began with D the application would be notified.
9.2 Sorting Data in selections generally needs to be sorted - often so that when displayed users can see data in a logical format. When a selection is defined the sorting order can be specified: the properties to sort on, in what order and what sorting algorithms to use.
9.3 Views There may be many items of data in a selection. Commonly when the data is being displayed it may not all fit on the screen and so the user will need to scroll up and down the data. A view provides this functionality by specifying the number of items of data the selection wants to deal with and the number of the first item of data out of the complete list of data the application wants to appear in the selection.
Views are important because they allow an application to limit the amount of data it stores locally to be limited to just the amount needed to display on the screen this reducing unnecessary duplication of data. 9.4 Efficiency
Cognima has made some efficiency optimisations in how the data is transferred between the Cognima server and client application — when multiple data changes are made the data is sent in blocks and then the application informed that the changes are complete so that the application only needs to update its user interface once.
9.5 Example
As an example we will define a selection called ContactSelection. This is the code that the framework will call back whenever a change is made to any of the selected objects. In the Cognima framework this is implemented as an object which you derive from the COdbSelect templated class - specifiying the type of object you want to have in the selection as the template argument, class CContactSelect : public COdbSelect<CContact>
{ public:
CContactSelect(COdb *aOdb); void ObjectAdded(CContact *aObject); void ObjectUpdated(CContact *aObject); void ObjectRemoved(const TOdbObjectld aObjectld); private: bool ListContactsO;
};
The methods ObjectAddedO, ObjectUpdatedO and ObjectRemovedO are called by the framework whenever respectively an object is added, updated or removed.
When you implement the Selection class you don't need to implement all these methods if you do not want to take instance action on any of these events - in some cases you may set up a selection to keep a list of a certain set of objects but only check that list on some other event and so the above methods would not be required. We have defined one extra private method called ListContactsO - this will list all the current contacts held by the selection. Here is the implementation of this class: CContactSelect::CContactSelect(COdb *aOdb) : COdbSelect<CContact>(aOdb)
{ }
void CContactSelect::ObjectAdded(CTestContact *aContact) {
OdbLog(OdbLogApp,L"New contact added: " « aContact->GetName()); ListContactsO;
}
void CContactSelect::ObjectUpdated(CTestContact *aContact)
{ OdbLog(OdbLogApp,L"Contact updated: " « aContact->GetName ); ListContactsO;
}
void CContactSelect::ObjectRemoved(const TOdbObjectld aObjecdd)
{ OdbLog(OdbLogApρ,L"Contact deleted - Id: " « aObjectld);
ListContactsO; }
void CContactSelecf.ListContactsO
{ OdbLog(OdbLogApp,L"Contacts list:");
for (unsigned long lndex=0; Index<iResult.size ; Index++) { CTestContact *Contact=iResult[Index] ; OdbLog(OdbLogApp,Contact->GetNameO « ", "
« Contact->GetPhoneO « ", "
<< Contact->GetEmailO); }
}
The constructor simply calls the default COdbSelect constructor. The ObjectAddedO, UpdatedO and RemovedO methods print out what change was made and then call ListContactsO to show what the current contents of the list is.
The ListContactsO shows how the current list of object held by the selection can be accessed. The current list of pointers to objects is held in a container class called iResult - this can then be accessed by normal container class integrators. In this case we simply go through the list and print all the objects in the list.

Claims

1. Method of automatically replicating data objects between a mobile device and a server, connected together via a wireless network, in which the timing of data replication across the network is determined by a network operator applying parameters that make efficient usage of network bandwidth.
2. The method of Claim 1 in which the parameters applied to a given object are both time dependent and also relate to how urgendy that object needs to be replicated.
3. The method of Claim 1 or 2 in which a change log lists all objects at the device and/or server to be replicated and the parameters then comprise a weight associated with each object that defines how urgently that object needs to be replicated.
4. The method of Claim 3 in which the parameters further comprise a threshold that is a function of time, with the weight of each object being locally compared to the threshold at a given time and the outcome of the comparison determining whether the object is sent for replication or not at that time.
5. The method of Claim 4 in which a connection is established at a given time if the weight of any object exceeds the threshold at that time.
6. The method of Claim 3 in which the weight of an object at a given time is a function of one or more of the following:
(a) Direction of data replication (device to server or server to device)
(b) Shelf life, defining the time or duration after which the object will be automatically deleted if still present in the change log
(c) Whether the object is overwri table (d) Size in bytes
(e) Time entered into the change log
(f) Priority
(g) Time out interval (h) Assigned time for replication
(i) User assignment of a non-default priority to a given object
(j) Memory available
7. The method of Claim 6 in which the network operator can cause the weight of an object to be altered at any time.
8. The method of Claim 4 in which the network operator can cause the threshold to be altered at any time.
9. The method of Claim 4 in which the threshold varies over time in such a way that efficient use is made of available bandwidth.
10. The method of Claim 4 in which the threshold can vary over time in a different way for different groups of end-users, individual end-users or applications.
11. The method of Claim 4 in which dynamic varying of the threshold can occur as cell or network loadings change.
12. The method of Claim 4 in which dynamic varying of the threshold can occur to encourage uptake of a new data replication service.
13. The method of Claim 4 in which the threshold can vary depending on one or more of the following:
(a) current time
(b) device roaming status
(c) cell or network loading
(d) time since last replication (e) user tariff
14. The method of Claim 4 in which, if the weight of no object exceeds the threshold at a given time, the time interval that will elapse before the weight of any object exceeds the threshold is calculated and a timer set for that time interval.
15. The method of Claim 14 in which the time interval is re-calculated if:
(a) a new object is added to the change log
(b) a new threshold is deployed
(c) the timer expires
(d) cell or network loading alters (e) device memory falls below a predefined level
(f) the device detects that its roaming state changes
(g) a new application is activated on the device (h) a network connection is terminated
16. The method of Claim 1 in which the end-user of the device can override default replication timing in respect of a specific object or type of object.
17. The method of Claim 1 in which an object to be replicated is assigned a time limit by which time replication must occur.
18. The method of Claim 17 in which the time limit is dynamic.
19. The method of Claim 17 in which the time limit alters if memory on the device changes or if the device roams to a new network
20. The method of Claim 1 in which an object to be replicated is assigned a shelf life which defines a time or duration after which the object will be deleted automatically if not replicated.
21. The method of Claim 1 in which different parameters enable the network operator to offer end-users different levels of data replication service, each associated with a different tariff.
22. The method of Claim 1 in which, once a connection initiating object has been replicated, then further objects in a change log and pending replication are sent as well.
23. The method of Claim 22 in which an opportunism threshold function determines the further objects that are sent.
24. The method of Claim 23 in which the opportunism threshold changes if the device is on a roaming network.
25. The method of Claim 23 in which the opportunism threshold changes depending on whether a cell loading threshold of the cell the device is located in is exceeded.
26. The method of Claim 23 in which the opportunism threshold is applied consistently by device and server, with changes to the threshold communicated across the network.
27. The method of Claim 23 in which the network operator can vary the opportunism threshold.
28. The method of Claim 1 in which the actual time of replication is a function of the state of the mobile device, the state of the network and the parameters.
29. A mobile device programmed with software that enables the device to replicate data to a server using the method of Claim 1.
30. A server programmed with software that enables the server to replicate data to a mobile device using the method of Claim 1.
PCT/GB2003/005598 2002-12-19 2003-12-19 Method of automatically replicating data objects between a mobile device and a server WO2004057828A1 (en)

Priority Applications (5)

Application Number Priority Date Filing Date Title
US10/539,863 US8156075B2 (en) 2002-12-19 2003-12-12 Method of automatically replicating data objects between a mobile device and a server
EP03782665.8A EP1576784B1 (en) 2002-12-19 2003-12-19 Method of automatically replicating data objects between a mobile device and a server
JP2004561666A JP2006511134A (en) 2002-12-19 2003-12-19 Method for automatically replicating data objects between a mobile device and a server
AU2003290306A AU2003290306A1 (en) 2002-12-19 2003-12-19 Method of automatically replicating data objects between a mobile device and a server
ES03782665.8T ES2689498T3 (en) 2002-12-19 2003-12-19 Method of automatic replication of data objects between a mobile device and a server

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
GB0229572.3 2002-12-19
GBGB0229572.3A GB0229572D0 (en) 2002-12-19 2002-12-19 Quality of service provisioning

Publications (1)

Publication Number Publication Date
WO2004057828A1 true WO2004057828A1 (en) 2004-07-08

Family

ID=9949975

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/GB2003/005598 WO2004057828A1 (en) 2002-12-19 2003-12-19 Method of automatically replicating data objects between a mobile device and a server

Country Status (7)

Country Link
US (1) US8156075B2 (en)
EP (1) EP1576784B1 (en)
JP (1) JP2006511134A (en)
AU (1) AU2003290306A1 (en)
ES (1) ES2689498T3 (en)
GB (2) GB0229572D0 (en)
WO (1) WO2004057828A1 (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1691521A1 (en) * 2005-02-09 2006-08-16 Deutsche Post AG Data transmission system, messaging component, and method of transmitting data
EP1805977A2 (en) * 2004-10-27 2007-07-11 Verisign, Inc. A method and apparatus for management of data on handheld
JP2007264685A (en) * 2006-03-27 2007-10-11 Fujitsu Ltd Mirror site operation program, method, and mirror site server computer
WO2009004276A1 (en) * 2007-07-02 2009-01-08 British Telecommunications Public Limited Company Method of synchronizing intermittently connected mobile terminals
WO2010010408A1 (en) 2008-07-25 2010-01-28 Vodafone Group Plc Mobile communications network
US8190923B2 (en) * 2005-12-20 2012-05-29 Microsoft Corporation Method to securely initialize, protect and recover system date/time
EP2353135A4 (en) * 2008-11-13 2012-11-28 Samsung Electronics Co Ltd System and method for providing a personalized mobile advertising service

Families Citing this family (70)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7305233B2 (en) * 2004-05-27 2007-12-04 Exclaim, Inc. Method and apparatus for image distribution using a cellular phone
GB2413654B (en) * 2004-04-29 2008-02-13 Symbian Software Ltd A method of backing up and restoring data in a computing device
US7921162B2 (en) * 2004-07-19 2011-04-05 The Go Daddy Group, Inc. Notification system and method for domain name registrars
SE0402505L (en) * 2004-10-14 2006-04-15 Faelt Comm Ab Device for a mobile telephone system
US9015263B2 (en) 2004-10-29 2015-04-21 Go Daddy Operating Company, LLC Domain name searching with reputation rating
US7596585B2 (en) * 2004-11-03 2009-09-29 Honeywell International Inc. Object replication using information quality of service
JP2006133955A (en) * 2004-11-04 2006-05-25 Nec Corp Backup system and method for data inside mobile communication terminal, and mobile communication terminal and backup device used therefor
US8626719B2 (en) * 2004-11-11 2014-01-07 Emc Corporation Methods of managing and accessing e-mail
US20070180151A1 (en) * 2005-09-20 2007-08-02 Honeywell International Inc. Model driven message processing
EP1984842B1 (en) 2006-02-01 2014-09-17 Coco Communications Corp. Congestion management and latency prediction in csma media
US20080134253A1 (en) * 2006-12-04 2008-06-05 Grant Romain Data upload and broadcast system
US8190561B1 (en) * 2006-12-06 2012-05-29 At&T Mobility Ii Llc LDAP replication priority queuing mechanism
US8068821B2 (en) * 2007-03-29 2011-11-29 Alcatel Lucent Method and apparatus for providing content to users using unicast and broadcast wireless networks
US8041780B2 (en) * 2007-03-29 2011-10-18 Alcatel Lucent Method and apparatus for dynamically pushing content over wireless networks
US8588750B2 (en) * 2007-03-31 2013-11-19 Alcatel Lucent Method and apparatus for providing interactive services to users using unicast and broadcast wireless networks
US7728636B2 (en) * 2007-08-14 2010-06-01 Qimonda Ag Clock signal synchronizing device with inherent duty-cycle correction capability
US8924344B2 (en) * 2007-11-28 2014-12-30 Microsoft Corporation User profile replication
US20090307060A1 (en) * 2008-06-09 2009-12-10 Merz Christopher J Methods and systems for determining a loyalty profile for a financial transaction cardholder
US8032633B2 (en) * 2008-06-10 2011-10-04 International Business Machines Corporation Computer-implemented method for implementing a requester-side autonomic governor using feedback loop information to dynamically adjust a resource threshold of a resource pool scheme
US8250212B2 (en) * 2008-06-10 2012-08-21 International Business Machines Corporation Requester-side autonomic governor
KR101690469B1 (en) 2008-07-15 2016-12-27 임머숀 코퍼레이션 Systems and methods for mapping message contents to virtual physical properties for vibrotactile messaging
US20110197139A1 (en) * 2008-11-05 2011-08-11 Hewlett-Packard Development Company. L.P. Displaying Personalized Information in a Handheld Device
US8898108B2 (en) * 2009-01-14 2014-11-25 Vmware, Inc. System and method for scheduling data storage replication over a network
US9026131B2 (en) 2009-07-21 2015-05-05 Modena Enterprises, Llc Systems and methods for associating contextual information and a contact entry with a communication originating from a geographic location
US9222798B2 (en) 2009-12-22 2015-12-29 Modena Enterprises, Llc Systems and methods for identifying an activity of a user based on a chronological order of detected movements of a computing device
US9215735B2 (en) 2010-03-03 2015-12-15 Modena Enterprises, Llc Systems and methods for initiating communications with contacts based on a communication specification
WO2011150968A1 (en) * 2010-06-02 2011-12-08 Malvacom Ab Communication method and device
US20120096070A1 (en) * 2010-10-15 2012-04-19 Shane Bryzak Web application framework remoting model api
US20120197848A1 (en) * 2011-01-28 2012-08-02 International Business Machines Corporation Validation of ingested data
US20120197902A1 (en) 2011-01-28 2012-08-02 International Business Machines Corporation Data ingest optimization
US8438137B2 (en) 2011-02-28 2013-05-07 Hewlett-Packard Development Company, L.P. Automatic selection of source or target deduplication
US20120226722A1 (en) 2011-03-02 2012-09-06 International Business Machines Corporation Systems, methods, and computer program products for secure multi-enterprise storage
US9002926B2 (en) 2011-04-22 2015-04-07 Go Daddy Operating Company, LLC Methods for suggesting domain names from a geographic location data
US8856896B1 (en) * 2011-06-24 2014-10-07 Amazon Technologies, Inc. Transparently updating user credentials
US8942854B2 (en) * 2011-11-28 2015-01-27 Kohler Co. System and method for identifying electrical devices in a power management system
US20130311597A1 (en) * 2012-05-16 2013-11-21 Apple Inc. Locally backed cloud-based storage
CN103546266A (en) * 2012-07-12 2014-01-29 三星电子株式会社 Smart scheduled sync method for sync applications
US20150234908A1 (en) * 2012-09-24 2015-08-20 President And Fellows Of Harvard College Techniques for data synchronization using compressive sensing
US9980114B2 (en) 2013-03-15 2018-05-22 Elwha Llc Systems and methods for communication management
US9451394B2 (en) 2012-12-31 2016-09-20 Elwha Llc Cost-effective mobile connectivity protocols
US9713013B2 (en) 2013-03-15 2017-07-18 Elwha Llc Protocols for providing wireless communications connectivity maps
US9832628B2 (en) 2012-12-31 2017-11-28 Elwha, Llc Cost-effective mobile connectivity protocols
US9781664B2 (en) 2012-12-31 2017-10-03 Elwha Llc Cost-effective mobile connectivity protocols
US9635605B2 (en) 2013-03-15 2017-04-25 Elwha Llc Protocols for facilitating broader access in wireless communications
US9876762B2 (en) 2012-12-31 2018-01-23 Elwha Llc Cost-effective mobile connectivity protocols
US9706382B2 (en) 2013-03-15 2017-07-11 Elwha Llc Protocols for allocating communication services cost in wireless communications
US9866706B2 (en) 2013-03-15 2018-01-09 Elwha Llc Protocols for facilitating broader access in wireless communications
US9596584B2 (en) 2013-03-15 2017-03-14 Elwha Llc Protocols for facilitating broader access in wireless communications by conditionally authorizing a charge to an account of a third party
US9693214B2 (en) * 2013-03-15 2017-06-27 Elwha Llc Protocols for facilitating broader access in wireless communications
US9781554B2 (en) 2013-03-15 2017-10-03 Elwha Llc Protocols for facilitating third party authorization for a rooted communication device in wireless communications
US9813887B2 (en) 2013-03-15 2017-11-07 Elwha Llc Protocols for facilitating broader access in wireless communications responsive to charge authorization statuses
US9843917B2 (en) 2013-03-15 2017-12-12 Elwha, Llc Protocols for facilitating charge-authorized connectivity in wireless communications
US9706060B2 (en) 2013-03-15 2017-07-11 Elwha Llc Protocols for facilitating broader access in wireless communications
US9807582B2 (en) 2013-03-15 2017-10-31 Elwha Llc Protocols for facilitating broader access in wireless communications
US9715694B2 (en) 2013-10-10 2017-07-25 Go Daddy Operating Company, LLC System and method for website personalization from survey data
US9684918B2 (en) 2013-10-10 2017-06-20 Go Daddy Operating Company, LLC System and method for candidate domain name generation
US10521309B1 (en) * 2013-12-23 2019-12-31 EMC IP Holding Company LLC Optimized filesystem walk for backup operations
US9992090B2 (en) * 2014-01-08 2018-06-05 Bank Of America Corporation Data metrics analytics
US9420222B2 (en) * 2014-01-28 2016-08-16 Nokia Technologies Oy Automatic image deletion
CN103825952B (en) * 2014-03-04 2017-07-04 百度在线网络技术(北京)有限公司 Cell dictionary method for pushing and server
US9838948B2 (en) * 2014-07-29 2017-12-05 Aruba Networks, Inc. Deep packet inspection (DPI) aware client steering and load balancing in wireless local area network (WLAN) infrastructure
US9953105B1 (en) 2014-10-01 2018-04-24 Go Daddy Operating Company, LLC System and method for creating subdomains or directories for a domain name
US9779125B2 (en) 2014-11-14 2017-10-03 Go Daddy Operating Company, LLC Ensuring accurate domain name contact information
US9785663B2 (en) 2014-11-14 2017-10-10 Go Daddy Operating Company, LLC Verifying a correspondence address for a registrant
US9953018B2 (en) * 2015-05-01 2018-04-24 Microsoft Technology Licensing, Llc Transfer of content between documents of different storage types
US10171314B2 (en) * 2015-12-01 2019-01-01 Here Global B.V. Methods, apparatuses and computer program products to derive quality data from an eventually consistent system
CA3015337C (en) * 2016-02-22 2022-01-25 Hubbell Incorporated Auto-adjusting data log record timestamps
US10783165B2 (en) * 2017-05-17 2020-09-22 International Business Machines Corporation Synchronizing multiple devices
US11048430B2 (en) 2019-04-12 2021-06-29 Netapp, Inc. Object store mirroring where during resync of two storage bucket, objects are transmitted to each of the two storage bucket
US12032593B2 (en) * 2022-11-21 2024-07-09 Bank Of America Corporation Real-time data replication monitoring and management

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0794646A2 (en) * 1996-03-08 1997-09-10 International Business Machines Corporation Data management system and method for replicated data
WO2001078319A2 (en) * 2000-04-10 2001-10-18 Research In Motion Limited System and method for bundling information

Family Cites Families (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5664175A (en) * 1993-08-13 1997-09-02 International Business Machines Corporation Method and system for reprioritizing calendar items on a data processing system
US6412017B1 (en) * 1996-07-01 2002-06-25 Microsoft Corporation Urgent replication facility
US6298386B1 (en) * 1996-08-14 2001-10-02 Emc Corporation Network file server having a message collector queue for connection and connectionless oriented protocols
US6442570B1 (en) * 1997-10-27 2002-08-27 Microsoft Corporation Object identification and data communication during an object synchronization process
US6779019B1 (en) * 1998-05-29 2004-08-17 Research In Motion Limited System and method for pushing information from a host system to a mobile data communication device
JP3578385B2 (en) * 1998-10-22 2004-10-20 インターナショナル・ビジネス・マシーンズ・コーポレーション Computer and replica identity maintaining method
GB2345824B (en) 1998-11-27 2001-05-23 Minorplanet Ltd Transferring operational data
US6865185B1 (en) * 2000-02-25 2005-03-08 Cisco Technology, Inc. Method and system for queuing traffic in a wireless communications network
US7307954B1 (en) * 2000-06-23 2007-12-11 Nokia Corporation Differentiated service network and method of operating a differentiated service network
AU2001278953A1 (en) * 2000-07-28 2002-02-13 American Calcar, Inc. Technique for effective organization and communication of information
US7185104B1 (en) * 2000-08-07 2007-02-27 At&T Corp. Methods and systems for optimizing network traffic
US6718348B1 (en) * 2000-08-25 2004-04-06 Telefonaktiebolaget Lm Ericsson (Publ) Non-time dependent synchronization of databases
US20020156921A1 (en) * 2001-04-19 2002-10-24 International Business Machines Corporation Automatic backup of wireless mobile device data onto gateway server while device is idle
US7032003B1 (en) * 2001-08-13 2006-04-18 Union Gold Holdings, Ltd. Hybrid replication scheme with data and actions for wireless devices
GB0128243D0 (en) 2001-11-26 2002-01-16 Cognima Ltd Cognima patent
US20030227487A1 (en) * 2002-06-01 2003-12-11 Hugh Harlan M. Method and apparatus for creating and accessing associative data structures under a shared model of categories, rules, triggers and data relationship permissions
US7023843B2 (en) * 2002-06-26 2006-04-04 Nokia Corporation Programmable scheduling for IP routers
US7318075B2 (en) * 2004-02-06 2008-01-08 Microsoft Corporation Enhanced tabular data stream protocol

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0794646A2 (en) * 1996-03-08 1997-09-10 International Business Machines Corporation Data management system and method for replicated data
WO2001078319A2 (en) * 2000-04-10 2001-10-18 Research In Motion Limited System and method for bundling information

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1805977A4 (en) * 2004-10-27 2009-04-22 Verisign Inc A method and apparatus for management of data on handheld
EP1805977A2 (en) * 2004-10-27 2007-07-11 Verisign, Inc. A method and apparatus for management of data on handheld
WO2006084500A1 (en) * 2005-02-09 2006-08-17 Deutsche Post Ag Data transmission system, a notification component and a data transmission method
EP1696625A1 (en) * 2005-02-09 2006-08-30 Deutsche Post AG Data transmission system, messaging component and method of transmitting data
EP1691521A1 (en) * 2005-02-09 2006-08-16 Deutsche Post AG Data transmission system, messaging component, and method of transmitting data
US8190923B2 (en) * 2005-12-20 2012-05-29 Microsoft Corporation Method to securely initialize, protect and recover system date/time
JP2007264685A (en) * 2006-03-27 2007-10-11 Fujitsu Ltd Mirror site operation program, method, and mirror site server computer
EP2028813A1 (en) * 2007-07-02 2009-02-25 British Telecmmunications public limited campany Method of synchronizing intermittently connected mobile terminals
WO2009004276A1 (en) * 2007-07-02 2009-01-08 British Telecommunications Public Limited Company Method of synchronizing intermittently connected mobile terminals
WO2010010408A1 (en) 2008-07-25 2010-01-28 Vodafone Group Plc Mobile communications network
GB2463000B (en) * 2008-07-25 2012-11-07 Vodafone Plc Mobile communications network
US8750203B2 (en) 2008-07-25 2014-06-10 Vodafone Group Plc Mobile communications network
US9247560B2 (en) 2008-07-25 2016-01-26 Vodafone Group Plc Mobile communications network
EP2353135A4 (en) * 2008-11-13 2012-11-28 Samsung Electronics Co Ltd System and method for providing a personalized mobile advertising service

Also Published As

Publication number Publication date
US8156075B2 (en) 2012-04-10
GB0229572D0 (en) 2003-01-22
AU2003290306A1 (en) 2004-07-14
EP1576784B1 (en) 2018-07-11
GB2397474A (en) 2004-07-21
EP1576784A1 (en) 2005-09-21
GB2397474B (en) 2005-06-15
JP2006511134A (en) 2006-03-30
ES2689498T3 (en) 2018-11-14
GB0329519D0 (en) 2004-01-28
US20060171523A1 (en) 2006-08-03

Similar Documents

Publication Publication Date Title
US8156075B2 (en) Method of automatically replicating data objects between a mobile device and a server
US20050108185A1 (en) Method of updating a display screen on a battery powered mobile computing device
US7894783B2 (en) Method of power management in a data replication process deployed in a wireless device
US9049212B2 (en) Method, system, and computer program product for prefetching sync data and for edge caching sync data on a cellular device
US10992617B2 (en) Transmitting rich media between a personal server and a mobile device over a wide area network
CA2216533C (en) Method and apparatus for rate governing communications
CN102207957B (en) Partial item change tracking and synchronization
US20070185899A1 (en) Likelihood-based storage management
US20050120050A1 (en) Enhanced storing of personal content
US20050289216A1 (en) Providing personalized services for mobile users
US20050215236A1 (en) Providing information for mobile users
CN101087262A (en) An instant communication processing method and system
GB2387687A (en) Method of replicating data between computing devices which use local clocks
US20040240632A1 (en) Generating messages on a mobile terminal using templates
KR100894705B1 (en) Mobile phone for managing individual information, and method for managing individual information on mobile phone, and a medium having computer readable program for executing the method
KR20070111725A (en) System and method for managing individual information, and a medium having computer readable program for executing the method

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NO NZ OM PH PL PT RO RU SD SE SG SK SL TJ TM TN TR TT TZ UA UG US UZ VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): BW GH GM KE LS MW MZ SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LU MC NL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
WWE Wipo information: entry into national phase

Ref document number: 2004561666

Country of ref document: JP

ENP Entry into the national phase

Ref document number: 2006171523

Country of ref document: US

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 10539863

Country of ref document: US

WWE Wipo information: entry into national phase

Ref document number: 2003782665

Country of ref document: EP

WWP Wipo information: published in national office

Ref document number: 2003782665

Country of ref document: EP

WWP Wipo information: published in national office

Ref document number: 10539863

Country of ref document: US