[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

US8572130B2 - Replacement policy for resource container - Google Patents

Replacement policy for resource container Download PDF

Info

Publication number
US8572130B2
US8572130B2 US13/169,925 US201113169925A US8572130B2 US 8572130 B2 US8572130 B2 US 8572130B2 US 201113169925 A US201113169925 A US 201113169925A US 8572130 B2 US8572130 B2 US 8572130B2
Authority
US
United States
Prior art keywords
resource
stored
queue
timestamp
resources
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US13/169,925
Other versions
US20120331019A1 (en
Inventor
Ivan Schreter
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
SAP SE
Original Assignee
SAP SE
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by SAP SE filed Critical SAP SE
Priority to US13/169,925 priority Critical patent/US8572130B2/en
Assigned to SAP AG reassignment SAP AG ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: SCHRETER, IVAN
Priority to EP12004744.4A priority patent/EP2541423B1/en
Publication of US20120331019A1 publication Critical patent/US20120331019A1/en
Priority to US14/065,002 priority patent/US8819074B2/en
Application granted granted Critical
Publication of US8572130B2 publication Critical patent/US8572130B2/en
Assigned to SAP SE reassignment SAP SE CHANGE OF NAME (SEE DOCUMENT FOR DETAILS). Assignors: SAP AG
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/08Addressing or allocation; Relocation in hierarchically structured memory systems, e.g. virtual memory systems
    • G06F12/12Replacement control
    • G06F12/121Replacement control using replacement algorithms
    • G06F12/123Replacement control using replacement algorithms with age lists, e.g. queue, most recently used [MRU] list or least recently used [LRU] list
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/22Indexing; Data structures therefor; Storage structures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/08Addressing or allocation; Relocation in hierarchically structured memory systems, e.g. virtual memory systems
    • G06F12/0802Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches
    • G06F12/0866Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches for peripheral storage systems, e.g. disk cache
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/22Indexing; Data structures therefor; Storage structures
    • G06F16/2282Tablespace storage structures; Management thereof
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/24Querying
    • G06F16/245Query processing
    • G06F16/2455Query execution
    • G06F16/24552Database cache management
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F5/00Methods or arrangements for data conversion without changing the order or content of the data handled
    • G06F5/06Methods or arrangements for data conversion without changing the order or content of the data handled for changing the speed of data flow, i.e. speed regularising or timing, e.g. delay lines, FIFO buffers; over- or underrun control therefor
    • G06F5/065Partitioned buffers, e.g. allowing multiple independent queues, bidirectional FIFO's
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2212/00Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
    • G06F2212/46Caching storage objects of specific type in disk cache
    • G06F2212/465Structured object, e.g. database record
    • G06F9/3855
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • G06F9/38Concurrent instruction execution, e.g. pipeline or look ahead
    • G06F9/3854Instruction completion, e.g. retiring, committing or graduating
    • G06F9/3856Reordering of instructions, e.g. using queues or age tags

Definitions

  • Some embodiments relate to database systems. In particular, some embodiments concern deallocation of cached resources within a database system.
  • a conventional database system uses a datastore to persistently store data pages, and a cache to provide fast access to the data pages. For example, in response to a request to access a data page, the data page is loaded from the datastore into the cache, and may thereafter be accessed from and/or modified within the cache. Once the data stored in the cache reaches a maximum allowed size, data pages are deallocated from the cache according to a replacement policy. For example, a replacement policy may result in deallocation of the oldest data pages in the cache until the data stored in the cache is of an acceptable size.
  • a modern in-memory database system uses volatile memory (e.g., Random Access Memory) for its cache and for persistent storage. Accordingly, it may be advantageous to selectively increase or decrease the cache size based on competing needs of the database system.
  • a replacement policy is also required in order to deallocate cached data pages so that the maximum cache size may be decreased to a desired level.
  • a least-recently used (LRU) policy may, for example, maintain a linked list of all cached data pages. The list is ordered based on the time at which each page was last used. For example, the least-recently used page is located at the end of the list, while the most-recently used page is located at the beginning of the list. If a command is received to free a particular amount of cache space, a number of data pages having a total size equal to or greater than the particular amount is identified from the end of the list and deallocated.
  • a conventional LRU policy therefore requires modification of the linked list in response to every page usage. Modification of the list requires an exclusive lock and therefore contention for the lock and resulting inefficiencies would be significant.
  • FIG. 1 is a block diagram of a database according to some embodiments.
  • FIG. 2 is a detailed block diagram of a database according to some embodiments.
  • FIG. 3 is a flow diagram of a process according to some embodiments.
  • FIGS. 4A through 4D illustrate a resource queue over time according to some embodiments.
  • FIGS. 5A and 5B comprise a flow diagram of a process according to some embodiments.
  • FIG. 1 is a block diagram of database system 100 according to some embodiments.
  • Database system 100 includes database server process 110 , cache 120 and datastore 130 .
  • Database server process 110 may comprise and/or may be implemented by computer-executable program code.
  • database system 100 may comprise one or more hardware devices, including at least one processor to execute program code so as to cause the one or more hardware devices to provide database server process 110 .
  • Cache 120 may be implemented in Random Access Memory (RAM), and datastore 130 may comprise one or more fixed disks.
  • database system 100 may comprise an “in-memory” database, in which both the data of datastore 130 and cache 120 are stored in volatile (e.g., non-disk-based) memory (e.g., RAM).
  • cache 120 may be capable of storing any objects of any size.
  • the stored objects may comprise one or more of conventional tabular data pages, row-based table data, columnar table data, dictionaries (e.g., for mapping of identifiers in columnar data to columnar tables), and object data.
  • Each stored object may be identified using a unique identifier. Pointers to each object may be stored in a lock-free hash table hashed by the identifiers in order to provide fast and scalable lookup.
  • Each cached object may also be associated with one or more properties, including but not limited to: a reference count, a flag indicating whether or not the page has been modified from its “last known good” version, and usage statistics/control information for a replacement policy.
  • properties may be maintained by a control block allocated in the cache for each cached object.
  • the control block may include any other information used for cache management, such as the logical page number of the cached object, or a read/write lock for the contents of the object.
  • the control block may implement a resource header, such that a single “resource” comprises a resource header and a resource proper (i.e., database data).
  • each user i.e., thread
  • the aforementioned reference count reflects a current number of users, and is decremented by each user upon release of the object. Accordingly, a cached object associated with a reference count of zero is not being used.
  • the replacement policies described below may utilize resource properties in addition to those mentioned above. These resource properties may include a timestamp indicating a time at which the resource was last-used, left and right child pointers to define a position of the resource in a resource queue, and/or a disposition of the resource to provide finer-grained control over the effective cache lifetimes of different types of cached objects.
  • Database system 100 may communicate with one or more database applications (not shown) over one or more interfaces (e.g., a Structured Query Language (SQL)-based interface).
  • the database applications may provide, for example, business reporting, inventory control, online shopping, and/or any other suitable functions.
  • the database applications may, in turn, support client applications that may be executed by client devices.
  • client application may simply comprise a Web browser to access and display reports generated by a database application.
  • the data of database system 100 may be received from disparate hardware and software systems, some of which are not interoperational with one another.
  • the systems may comprise a back-end data environment employed in a business or industrial context.
  • the data may be pushed to database system 100 and/or provided in response to queries received therefrom.
  • Database system 100 and each element thereof may also include other unshown elements that may be used during operation thereof, such as any suitable program code, scripts, or other functional data that is executable to interface with other elements, other applications, other data files, operating system files, and device drivers. These elements are known to those in the art, and are therefore not described in detail herein.
  • database system 100 which is a “single node” database system
  • embodiments may also be implemented within one or more nodes of a distributed database, each of which comprises an executing process, a cache and a datastore.
  • the data stored in the datastores of each node, taken together, represent the full database, and the database server processes of each node operate to transparently provide the data of the full database to the aforementioned database applications.
  • Database system 100 may also or alternatively support multi-tenancy by providing multiple logical database systems which are programmatically isolated from one another.
  • FIG. 2 is a block diagram of database system 100 according to some embodiments.
  • datastore 130 comprises only data volume 132 .
  • Datastore 130 may comprise one or more data volumes in some embodiments, with each of the one or more data volumes comprising one or more disparate physical systems for storing data. These physical systems may comprise a portion of a physical hard disk, an entire physical hard disk, a storage system composed of several physical hard disks, and/or RAM.
  • a data volume is subdivided into storage areas known as blocks, and data is stored in the data volume in data pages 1322 , each of which has the same size as a block. Accordingly, a particular data page 1322 of datastore 130 may be accessed by referencing the data volume and block address associated with that data page.
  • Data pages 1322 may include application data consisting of tabular data, row-based data, columnar data, object data and associated index entries. In a case that datastore 130 includes more than one data volume, data pages 1322 may be spread across one or more of its data volumes.
  • Data volume 132 includes file directory 1324 and converter 1326 . If datastore 130 includes more than one data volume, file directory 1324 and converter 1326 may be spread across one or more of the data volumes. When a new data page is created, the data page is assigned a unique logical page number. Converter 1326 maps this logical page number to the data volume and block address at which the data page is stored.
  • File directory 1324 maps a file identifier to a logical page number of a corresponding file root page, and the aforementioned database catalog maps each file identifier to associated metadata, including a name of a database object associated with the file identifier. Accordingly, the information of the database catalog and file directory 1324 may be used to determine a logical page number from a name of a database object, for example.
  • Datastore 130 may also include configuration files 134 defining properties of database system 110 (e.g., a size and physical location of each data volume, a maximum number of data volumes in datastore 130 , etc.). Moreover, datastore 130 typically includes system files, database parameters, paths, user information and any other suitable information. Datastore 130 may also store a database catalog including metadata describing the database objects that are stored therein.
  • configuration files 134 defining properties of database system 110 (e.g., a size and physical location of each data volume, a maximum number of data volumes in datastore 130 , etc.).
  • datastore 130 typically includes system files, database parameters, paths, user information and any other suitable information.
  • Datastore 130 may also store a database catalog including metadata describing the database objects that are stored therein.
  • DB server process 110 may comprise any system for managing a database system that is or becomes known. Generally, DB server process 110 may receive requests for data (e.g., SQL requests from a database application), may retrieve the requested data from datastore 130 or from cache 120 , and may return the requested data to the requestor. In some embodiments, DB server process 110 includes SQL manager 112 to process received SQL statements and data access manager 114 to manage access to stored data. DB server process 110 may also perform start-up, logging, recovery, management, optimization, monitoring, indexing, integrity checks and other database-related tasks.
  • requests for data e.g., SQL requests from a database application
  • DB server process 110 includes SQL manager 112 to process received SQL statements and data access manager 114 to manage access to stored data.
  • DB server process 110 may also perform start-up, logging, recovery, management, optimization, monitoring, indexing, integrity checks and other database-related tasks.
  • Cache 120 stores various elements of datastore 130 during execution of database system 100 . These elements may include recently-accessed data 1322 , pages of converter 1326 , and/or a log queue. As shown, cache 120 includes converter 122 and cached resources 124 . Converter 122 and cached resources 124 are illustrated separately herein for the sake of clarity. However, according to some embodiments, converter 122 and cached resources 124 might not comprise separate, contiguous memory addresses of cache 120 . For example, pages of converter 122 may be interspersed among cached resources 124 throughout cache 120 .
  • the pages of converter 1326 are created in cache 120 as converter 122 in order to provide fast access to cached resources 1322 .
  • a database transaction may require access to a database object stored in cached resources 1322 .
  • a file identifier corresponding to the database object is determined based on the database catalog, and a logical page number of a root page of the database object is determined from the file identifier using file directory 1324 . If the root page is not already among cached resources 124 of cache 120 , converter 122 is traversed to determine a block address of data volume 132 at which the root page is stored, based on the logical page number.
  • the data page is read from the block address of data volume 132 and is created within cached resources 124 of cache 120 . If the data page is modified within cache 120 , a log entry describing the modification is recorded, the modified page remains in cache 120 , the modified page is designated as modified using the above-mentioned resource properties, and the original “last known good” page remains at the block address of data volume 132 from which the data page was read.
  • converter 1326 of data volume 132 maps logical page numbers to block addresses of data volume 132
  • converter 1326 must be modified once a corresponding data page is saved to a new location of data volume 132 .
  • the modified converter pages are flushed to data volume 132 at the end of a savepoint, particularly after all modified data pages are written to data volume 132 .
  • a restart record is created to point to the starting point of the newly-saved converter within data volume 132 .
  • the restart record may be stored in any volume of datastore 130 .
  • FIG. 3 comprises a flow diagram of process 300 according to some embodiments.
  • various hardware elements of a database execute program code to perform process 300 .
  • Process 300 may be performed by DB server process 110 according to some embodiments.
  • Process 300 and all other processes mentioned herein may be embodied in computer-executable program code read from one or more of non-transitory computer-readable media, such as a floppy disk, a CD-ROM, a DVD-ROM, a Flash drive, and a magnetic tape, and then stored in a compressed, uncompiled and/or encrypted format.
  • non-transitory computer-readable media such as a floppy disk, a CD-ROM, a DVD-ROM, a Flash drive, and a magnetic tape
  • hard-wired circuitry may be used in place of, or in combination with, program code for implementation of processes according to some embodiments. Embodiments are therefore not limited to any specific combination of hardware and software.
  • a resource queue is initially created at S 302 .
  • the resource queue indicates an order of cached resources that is based on respective timestamps associated with the cached resources.
  • each resource (or data page) stored in cache 120 is associated with a timestamp indicating a time at which the resource was last-used.
  • the timestamp need not be an actual time (e.g., 13:45:02.56 on Jun. 8, 2011), but may comprise any data which is usable to order resources based their relative times of last usage.
  • the timestamp for a resource may simply comprise a value of a continuously-incrementing counter register which is evaluated when the resource's reference count reaches zero.
  • the resource queue suggests a tree structure as shown in FIG. 4A .
  • Queue 400 represents each resource by its unique identifier, and each resource (except the lowest-level resources R 42 , R 3212 , R 455 and R 19 ) is associated with a pointer to a lower-left resource in the tree structure and a pointer to a lower-right resource in the tree structure. These two pointers may be defined by the resource header of the resource.
  • queue 400 might not exist as an independent data structure, instead it may be an abstract logical entity defined by the pointers of the corresponding resource headers.
  • resource queue 400 may indicate an order of resources based on their timestamps.
  • the resources of queue 400 are arranged such that a resource is initially associated with an earlier timestamp than all resources located to the right of the resource.
  • resource R 42 is associated with the earliest (i.e, oldest) timestamp and R 19 is associated with the most-recent timestamp.
  • the timestamp associated with resource R 1234 is older then the timestamps associated with resources R 455 and R 4552 , but more-recent than the timestamps associated with resources R 323 and R 3212 .
  • a timestamp associated with resource R 1234 may indicate a later time than the timestamps associated with resources R 323 and R 4522 .
  • the timestamp associated with resource R 323 may indicate a later time than the timestamps associated with resources R 42 and R 3212 .
  • the resource queue itself is timestamped at S 304 .
  • the resource queue timestamp may comprise any data that allows temporal comparison with the timestamps of the cached resources.
  • the command may comprise an instruction to reduce the utilization of the cache to a particular percentage of its total capacity, or to deallocate at least a particular percentage or number of bits.
  • the received command may comprise an instruction to shrink the total cache size by a particular percentage or a particular amount.
  • any cached resources specified in the queue are older than the queue.
  • one or more resources of the queue may have been used during the time between creation of the resource queue and reception of the command at S 306 , with their timestamps being updated accordingly.
  • S 308 determines whether any of the resources of the queue are associated with timestamps that are earlier than the queue's timestamp. If not, flow returns to S 302 to create a new resource queue as described above.
  • pointers queue 400 of FIG. 4A may be traversed quickly from its root to its lowest level to identify resource R 42 , which is assumed in the present example to be the oldest cached resource which is older than queue 400 .
  • FIG. 4B illustrates deallocation of resource R 42 at S 310 according to some embodiments. The pointers of parent resource R 323 remain intact to maintain the integrity of resource queue 400 , although resource R 42 is no longer allocated within the cache.
  • a timestamp associated with resource R 323 has been updated during execution of process 300 due to usage of resource R 323 sometime after S 304 . Accordingly, R 3212 , which may have otherwise been the oldest resource in queue 400 , is not deallocated at S 310 . Instead, resource R 3212 is deallocated. Again, the pointers of parent resource R 323 remain intact to maintain the integrity of resource queue 400 .
  • FIG. 4D shows queue 400 after execution of several of such cycles.
  • the resource header of resource R 323 has been updated during these cycles to indicate that resource R 323 has been modified (e.g., as indicated by *), and to update its timestamp accordingly.
  • Resources R 1234 and R 4522 have been deallocated, and the timestamp associated with resource R 455 has been updated to reflect a recent use.
  • resource R 19 may be deallocated in response to a next-received command.
  • queue 400 is not updated or regenerated (i.e., flow does not return to S 302 ) until all the resources of the queue have been deallocated or until all currently-cached resources specified in the queue are associated with timestamps newer than the queue's timestamp. As a result, the number and frequency of locks required to implement a replacement policy may be reduced.
  • Process 300 treats all types of resources equally. However, in some implementations, certain types of resources may be more desirable to maintain in a cache than others. For example, in an in-memory database system, it may be more important to cache in-memory table data for a columnar engine than to cache a database page.
  • a replacement policy may be based on a resource disposition assigned to each resource.
  • resource dispositions may include temporary, permanent, short-term and long-term.
  • resources having different resource dispositions may be treated differently by the replacement policy.
  • a resource associated with the temporary resource disposition may be automatically removed from the cache when its reference count reaches zero. If such a resource was modified in the cache, it is first flushed to persistent storage prior to removal.
  • Examples of resources which may be associated with the temporary resource disposition include database pages storing columnar data after a columnar merge, since they are not required for query processing after the merge, only for re-loading the table.
  • a resource associated with the permanent resource disposition is never removed by a replacement policy and is not included in a resource queue created as described herein. Such a resource must be removed from the cache explicitly.
  • Database pages which are used to store in-memory row data may be assigned the permanent resource disposition, since such pages are accessed using memory pointers and therefore cannot be simply removed.
  • Process 500 implements a replacement policy which supports more than one type of resource disposition.
  • Process 500 will be described with respect to an example including two resource dispositions, but embodiments are not limited thereto.
  • one resource queue is created for each resource disposition.
  • Each resource queue is then time-stamped and a “resource present” flag for each resource queue is set.
  • cached resources of each resource disposition are ordered within their corresponding resource queue based on their respective times since last use. This order is determined based on the associated timestamps. Therefore, in the present example, two resource queues exist after completion of S 506 , with the first resource queue indicating an order of cached resources of a first resource type, and with the second resource queue indicating an order of cached resources of a second resource type.
  • Each queue may suggest a tree structure using pointers as described above with respect to process 300 .
  • the resource-present flag of each queue is reset at S 508 .
  • the resource-present flag of a queue is set if the reference count associated with the resource reaches zero due to its release by a using thread. Updating in this sense also includes caching of a new resource associated with the resource disposition.
  • a weighted not-in-use time is determined for the oldest resources of each queue. For example, a weight of 10 may be assigned to resources of a short-term disposition queue, while resources of a long-term disposition queue may be assigned a weight of 1. Moreover, a not-in-use time may be determined for the oldest cached resource in each queue by subtracting the timestamp of the oldest cached resource from the current timestamp. The determined not-in-use time of each oldest cached resource is then multiplied by the weight assigned to its resource disposition to determine its weighted not-in-use time.
  • the oldest cached resource which is associated with the largest weighted not-in-use time is deallocated.
  • the short term resource associated with the weighted not-in-use time of 50 s is deallocated at S 522 .
  • the pointers of the resource's header, as well as any pointers to the resource, remain intact to maintain the integrity of its resource queue.
  • Process 500 therefore reduces the number and frequency of locks required to implement a replacement policy, while also providing a policy which may consider the relative priority of different types of resources.
  • Elements described herein as communicating with one another are directly or indirectly capable of communicating over any number of different systems for transferring data, including but not limited to shared memory communication, a local area network, a wide area network, a telephone network, a cellular network, a fiber-optic network, a satellite network, an infrared network, a radio frequency network, and any other type of network that may be used to transmit information between devices.
  • communication between systems may proceed over any one or more transmission protocols that are or become known, such as Asynchronous Transfer Mode (ATM), Internet Protocol (IP), Hypertext Transfer Protocol (HTTP) and Wireless Application Protocol (WAP).
  • ATM Asynchronous Transfer Mode
  • IP Internet Protocol
  • HTTP Hypertext Transfer Protocol
  • WAP Wireless Application Protocol

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Databases & Information Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Software Systems (AREA)
  • Computational Linguistics (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

A system includes creation of a first resource queue indicating an order of resources stored in a memory, the order based on respective timestamps associated with the stored resources, association of the first resource queue with a first queue timestamp, reception of a first command to deallocate a first amount of stored resources from the memory, determination that a first stored resource indicated by the first resource queue is associated with a timestamp earlier than the first queue timestamp, deallocation of the first stored resource from the memory, reception of a second command to deallocate a second amount of stored resources from the memory, determination that the first resource queue indicates no stored resources which are associated with a timestamp earlier than the first queue timestamp, and, in response to the determination that the first resource queue indicates no stored resources which are associated with a timestamp earlier than the first queue timestamp, creation of a second resource queue indicating a second order of second resources stored in the memory, the second order based on respective second timestamps associated with the stored second resources.

Description

FIELD
Some embodiments relate to database systems. In particular, some embodiments concern deallocation of cached resources within a database system.
BACKGROUND
A conventional database system uses a datastore to persistently store data pages, and a cache to provide fast access to the data pages. For example, in response to a request to access a data page, the data page is loaded from the datastore into the cache, and may thereafter be accessed from and/or modified within the cache. Once the data stored in the cache reaches a maximum allowed size, data pages are deallocated from the cache according to a replacement policy. For example, a replacement policy may result in deallocation of the oldest data pages in the cache until the data stored in the cache is of an acceptable size.
A modern in-memory database system uses volatile memory (e.g., Random Access Memory) for its cache and for persistent storage. Accordingly, it may be advantageous to selectively increase or decrease the cache size based on competing needs of the database system. A replacement policy is also required in order to deallocate cached data pages so that the maximum cache size may be decreased to a desired level.
Many types of replacement policies exist. A least-recently used (LRU) policy may, for example, maintain a linked list of all cached data pages. The list is ordered based on the time at which each page was last used. For example, the least-recently used page is located at the end of the list, while the most-recently used page is located at the beginning of the list. If a command is received to free a particular amount of cache space, a number of data pages having a total size equal to or greater than the particular amount is identified from the end of the list and deallocated.
A conventional LRU policy therefore requires modification of the linked list in response to every page usage. Modification of the list requires an exclusive lock and therefore contention for the lock and resulting inefficiencies would be significant.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is a block diagram of a database according to some embodiments.
FIG. 2 is a detailed block diagram of a database according to some embodiments.
FIG. 3 is a flow diagram of a process according to some embodiments.
FIGS. 4A through 4D illustrate a resource queue over time according to some embodiments.
FIGS. 5A and 5B comprise a flow diagram of a process according to some embodiments.
DETAILED DESCRIPTION
FIG. 1 is a block diagram of database system 100 according to some embodiments. Database system 100 includes database server process 110, cache 120 and datastore 130.
Database server process 110 may comprise and/or may be implemented by computer-executable program code. For example, database system 100 may comprise one or more hardware devices, including at least one processor to execute program code so as to cause the one or more hardware devices to provide database server process 110.
Cache 120 may be implemented in Random Access Memory (RAM), and datastore 130 may comprise one or more fixed disks. Alternatively, database system 100 may comprise an “in-memory” database, in which both the data of datastore 130 and cache 120 are stored in volatile (e.g., non-disk-based) memory (e.g., RAM).
In contrast to conventional fixed-size caches which store data pages of a fixed size, cache 120 may be capable of storing any objects of any size. In some embodiments, the stored objects may comprise one or more of conventional tabular data pages, row-based table data, columnar table data, dictionaries (e.g., for mapping of identifiers in columnar data to columnar tables), and object data. Each stored object may be identified using a unique identifier. Pointers to each object may be stored in a lock-free hash table hashed by the identifiers in order to provide fast and scalable lookup.
Each cached object may also be associated with one or more properties, including but not limited to: a reference count, a flag indicating whether or not the page has been modified from its “last known good” version, and usage statistics/control information for a replacement policy. Such properties may be maintained by a control block allocated in the cache for each cached object. The control block may include any other information used for cache management, such as the logical page number of the cached object, or a read/write lock for the contents of the object. Logically, the control block may implement a resource header, such that a single “resource” comprises a resource header and a resource proper (i.e., database data).
When a cached object is in-use, each user (i.e., thread) using the object holds a handle to the cached object. The aforementioned reference count reflects a current number of users, and is decremented by each user upon release of the object. Accordingly, a cached object associated with a reference count of zero is not being used.
The replacement policies described below may utilize resource properties in addition to those mentioned above. These resource properties may include a timestamp indicating a time at which the resource was last-used, left and right child pointers to define a position of the resource in a resource queue, and/or a disposition of the resource to provide finer-grained control over the effective cache lifetimes of different types of cached objects.
Database system 100 may communicate with one or more database applications (not shown) over one or more interfaces (e.g., a Structured Query Language (SQL)-based interface). The database applications may provide, for example, business reporting, inventory control, online shopping, and/or any other suitable functions. The database applications may, in turn, support client applications that may be executed by client devices. Such a client application may simply comprise a Web browser to access and display reports generated by a database application.
The data of database system 100 may be received from disparate hardware and software systems, some of which are not interoperational with one another. The systems may comprise a back-end data environment employed in a business or industrial context. The data may be pushed to database system 100 and/or provided in response to queries received therefrom.
Database system 100 and each element thereof may also include other unshown elements that may be used during operation thereof, such as any suitable program code, scripts, or other functional data that is executable to interface with other elements, other applications, other data files, operating system files, and device drivers. These elements are known to those in the art, and are therefore not described in detail herein.
Although embodiments are described with respect to database system 100, which is a “single node” database system, embodiments may also be implemented within one or more nodes of a distributed database, each of which comprises an executing process, a cache and a datastore. The data stored in the datastores of each node, taken together, represent the full database, and the database server processes of each node operate to transparently provide the data of the full database to the aforementioned database applications. Database system 100 may also or alternatively support multi-tenancy by providing multiple logical database systems which are programmatically isolated from one another.
FIG. 2 is a block diagram of database system 100 according to some embodiments. For purposes of the foregoing description, it will be assumed that datastore 130 comprises only data volume 132. Datastore 130 may comprise one or more data volumes in some embodiments, with each of the one or more data volumes comprising one or more disparate physical systems for storing data. These physical systems may comprise a portion of a physical hard disk, an entire physical hard disk, a storage system composed of several physical hard disks, and/or RAM.
Generally, a data volume is subdivided into storage areas known as blocks, and data is stored in the data volume in data pages 1322, each of which has the same size as a block. Accordingly, a particular data page 1322 of datastore 130 may be accessed by referencing the data volume and block address associated with that data page. Data pages 1322 may include application data consisting of tabular data, row-based data, columnar data, object data and associated index entries. In a case that datastore 130 includes more than one data volume, data pages 1322 may be spread across one or more of its data volumes.
Data volume 132 includes file directory 1324 and converter 1326. If datastore 130 includes more than one data volume, file directory 1324 and converter 1326 may be spread across one or more of the data volumes. When a new data page is created, the data page is assigned a unique logical page number. Converter 1326 maps this logical page number to the data volume and block address at which the data page is stored. File directory 1324 maps a file identifier to a logical page number of a corresponding file root page, and the aforementioned database catalog maps each file identifier to associated metadata, including a name of a database object associated with the file identifier. Accordingly, the information of the database catalog and file directory 1324 may be used to determine a logical page number from a name of a database object, for example.
The foregoing process also applies to “in-memory” implementations. However, an identifier of a data volume in which a page is stored might not be utilized in such implementations, as the in-memory datastore might simply comprise addressable memory locations which are not divided into logical data volumes.
Datastore 130 may also include configuration files 134 defining properties of database system 110 (e.g., a size and physical location of each data volume, a maximum number of data volumes in datastore 130, etc.). Moreover, datastore 130 typically includes system files, database parameters, paths, user information and any other suitable information. Datastore 130 may also store a database catalog including metadata describing the database objects that are stored therein.
DB server process 110 may comprise any system for managing a database system that is or becomes known. Generally, DB server process 110 may receive requests for data (e.g., SQL requests from a database application), may retrieve the requested data from datastore 130 or from cache 120, and may return the requested data to the requestor. In some embodiments, DB server process 110 includes SQL manager 112 to process received SQL statements and data access manager 114 to manage access to stored data. DB server process 110 may also perform start-up, logging, recovery, management, optimization, monitoring, indexing, integrity checks and other database-related tasks.
Cache 120 stores various elements of datastore 130 during execution of database system 100. These elements may include recently-accessed data 1322, pages of converter 1326, and/or a log queue. As shown, cache 120 includes converter 122 and cached resources 124. Converter 122 and cached resources 124 are illustrated separately herein for the sake of clarity. However, according to some embodiments, converter 122 and cached resources 124 might not comprise separate, contiguous memory addresses of cache 120. For example, pages of converter 122 may be interspersed among cached resources 124 throughout cache 120.
The pages of converter 1326 are created in cache 120 as converter 122 in order to provide fast access to cached resources 1322. In this regard, a database transaction may require access to a database object stored in cached resources 1322. A file identifier corresponding to the database object is determined based on the database catalog, and a logical page number of a root page of the database object is determined from the file identifier using file directory 1324. If the root page is not already among cached resources 124 of cache 120, converter 122 is traversed to determine a block address of data volume 132 at which the root page is stored, based on the logical page number.
The data page is read from the block address of data volume 132 and is created within cached resources 124 of cache 120. If the data page is modified within cache 120, a log entry describing the modification is recorded, the modified page remains in cache 120, the modified page is designated as modified using the above-mentioned resource properties, and the original “last known good” page remains at the block address of data volume 132 from which the data page was read.
Once the number of modified pages in cache 120 reaches a threshold amount, or after passage of a designated time interval, all pages of cache 120 which are designated as modified are written to data volume 132. The modified page is not written to the block address of data volume 132 from which it was initially read. Rather, the original unmodified page remains designated as a “last known good” page at its block address and the modified page is written to a new block address of data volume 132.
Since converter 1326 of data volume 132 maps logical page numbers to block addresses of data volume 132, converter 1326 must be modified once a corresponding data page is saved to a new location of data volume 132. The modified converter pages are flushed to data volume 132 at the end of a savepoint, particularly after all modified data pages are written to data volume 132. Then, a restart record is created to point to the starting point of the newly-saved converter within data volume 132. The restart record may be stored in any volume of datastore 130.
FIG. 3 comprises a flow diagram of process 300 according to some embodiments. In some embodiments, various hardware elements of a database execute program code to perform process 300. Process 300 may be performed by DB server process 110 according to some embodiments.
Process 300 and all other processes mentioned herein may be embodied in computer-executable program code read from one or more of non-transitory computer-readable media, such as a floppy disk, a CD-ROM, a DVD-ROM, a Flash drive, and a magnetic tape, and then stored in a compressed, uncompiled and/or encrypted format. In some embodiments, hard-wired circuitry may be used in place of, or in combination with, program code for implementation of processes according to some embodiments. Embodiments are therefore not limited to any specific combination of hardware and software.
A resource queue is initially created at S302. The resource queue indicates an order of cached resources that is based on respective timestamps associated with the cached resources. According to some embodiments, each resource (or data page) stored in cache 120 is associated with a timestamp indicating a time at which the resource was last-used. The timestamp need not be an actual time (e.g., 13:45:02.56 on Jun. 8, 2011), but may comprise any data which is usable to order resources based their relative times of last usage. For example, the timestamp for a resource may simply comprise a value of a continuously-incrementing counter register which is evaluated when the resource's reference count reaches zero.
According to some embodiments, the resource queue suggests a tree structure as shown in FIG. 4A. Queue 400 represents each resource by its unique identifier, and each resource (except the lowest-level resources R42, R3212, R455 and R19) is associated with a pointer to a lower-left resource in the tree structure and a pointer to a lower-right resource in the tree structure. These two pointers may be defined by the resource header of the resource. In this regard, queue 400 might not exist as an independent data structure, instead it may be an abstract logical entity defined by the pointers of the corresponding resource headers.
As mentioned, resource queue 400 may indicate an order of resources based on their timestamps. According to the embodiment described below, the resources of queue 400 are arranged such that a resource is initially associated with an earlier timestamp than all resources located to the right of the resource. For example, resource R42 is associated with the earliest (i.e, oldest) timestamp and R19 is associated with the most-recent timestamp. Also, the timestamp associated with resource R1234 is older then the timestamps associated with resources R455 and R4552, but more-recent than the timestamps associated with resources R323 and R3212.
In other examples, a timestamp associated with resource R1234 may indicate a later time than the timestamps associated with resources R323 and R4522. Similarly, the timestamp associated with resource R323 may indicate a later time than the timestamps associated with resources R42 and R3212. Some embodiments may reverse this ordering such that resources associated with earlier timestamps are located higher in the tree structure.
The resource queue itself is timestamped at S304. The resource queue timestamp may comprise any data that allows temporal comparison with the timestamps of the cached resources.
Flow cycles at S306 until a command is received to deallocate a particular amount of resources from the cache. In the case of a fixed-size cache, the command may comprise an instruction to reduce the utilization of the cache to a particular percentage of its total capacity, or to deallocate at least a particular percentage or number of bits. In the case of a variable-sized cache, the received command may comprise an instruction to shrink the total cache size by a particular percentage or a particular amount.
After the command is received, it is determined, at S308, whether any cached resources specified in the queue are older than the queue. In this regard, one or more resources of the queue may have been used during the time between creation of the resource queue and reception of the command at S306, with their timestamps being updated accordingly. S308 determines whether any of the resources of the queue are associated with timestamps that are earlier than the queue's timestamp. If not, flow returns to S302 to create a new resource queue as described above.
If the determination at S308 is affirmative, flow proceeds to S310 to deallocate the oldest cached resource of the queue which is older than the queue. Using the above-mentioned pointers queue 400 of FIG. 4A may be traversed quickly from its root to its lowest level to identify resource R42, which is assumed in the present example to be the oldest cached resource which is older than queue 400. FIG. 4B illustrates deallocation of resource R42 at S310 according to some embodiments. The pointers of parent resource R323 remain intact to maintain the integrity of resource queue 400, although resource R42 is no longer allocated within the cache.
Flow continues to S312 to determine whether the required amount of resources has been deallocated. If not, flow returns to S308 to determine whether any cached resources of the queue are older than the queue, as described above. Assuming an affirmative determination, flow again arrives at S310 to deallocate the oldest cached resource of the queue which is older than the queue.
As shown in FIG. 4C, a timestamp associated with resource R323 has been updated during execution of process 300 due to usage of resource R323 sometime after S304. Accordingly, R3212, which may have otherwise been the oldest resource in queue 400, is not deallocated at S310. Instead, resource R3212 is deallocated. Again, the pointers of parent resource R323 remain intact to maintain the integrity of resource queue 400.
Flow therefore cycles between S308, S310 and S312 until the required amount of resources has been deallocated from the queue. FIG. 4D shows queue 400 after execution of several of such cycles. The resource header of resource R323 has been updated during these cycles to indicate that resource R323 has been modified (e.g., as indicated by *), and to update its timestamp accordingly. Resources R1234 and R4522 have been deallocated, and the timestamp associated with resource R455 has been updated to reflect a recent use.
It will be assumed that the required amount of resources is then determined to have been deallocated at S312. Flow therefore returns to S306 to wait for another command to deallocate an amount of resources from the cache. If such a command is received, flow proceeds to step S308 and continues as described above. With respect to the example of FIG. 4D, resource R19 may be deallocated in response to a next-received command. Notably, queue 400 is not updated or regenerated (i.e., flow does not return to S302) until all the resources of the queue have been deallocated or until all currently-cached resources specified in the queue are associated with timestamps newer than the queue's timestamp. As a result, the number and frequency of locks required to implement a replacement policy may be reduced.
Process 300 treats all types of resources equally. However, in some implementations, certain types of resources may be more desirable to maintain in a cache than others. For example, in an in-memory database system, it may be more important to cache in-memory table data for a columnar engine than to cache a database page.
A replacement policy may be based on a resource disposition assigned to each resource. Examples of resource dispositions may include temporary, permanent, short-term and long-term. Generally, resources having different resource dispositions may be treated differently by the replacement policy.
A resource associated with the temporary resource disposition may be automatically removed from the cache when its reference count reaches zero. If such a resource was modified in the cache, it is first flushed to persistent storage prior to removal. Examples of resources which may be associated with the temporary resource disposition include database pages storing columnar data after a columnar merge, since they are not required for query processing after the merge, only for re-loading the table.
A resource associated with the permanent resource disposition is never removed by a replacement policy and is not included in a resource queue created as described herein. Such a resource must be removed from the cache explicitly. Database pages which are used to store in-memory row data (e.g., containing columnar table metadata) may be assigned the permanent resource disposition, since such pages are accessed using memory pointers and therefore cannot be simply removed.
Process 500 implements a replacement policy which supports more than one type of resource disposition. Process 500 will be described with respect to an example including two resource dispositions, but embodiments are not limited thereto. Initially, at S502, one resource queue is created for each resource disposition. Each resource queue is then time-stamped and a “resource present” flag for each resource queue is set.
Next, at S506, cached resources of each resource disposition are ordered within their corresponding resource queue based on their respective times since last use. This order is determined based on the associated timestamps. Therefore, in the present example, two resource queues exist after completion of S506, with the first resource queue indicating an order of cached resources of a first resource type, and with the second resource queue indicating an order of cached resources of a second resource type. Each queue may suggest a tree structure using pointers as described above with respect to process 300.
The resource-present flag of each queue is reset at S508. During subsequent execution of process 500, the resource-present flag of a queue is set if the reference count associated with the resource reaches zero due to its release by a using thread. Updating in this sense also includes caching of a new resource associated with the resource disposition.
Flow cycles at S510 until a command is received to deallocate a particular amount of resources from the cache. After the command is received, it is determined, at S512, whether any of the queues are empty (i.e., all resources of the queue have been deallocated) and associated with a set resource-present flag (i.e., the cache includes one or more resources associated with the resource disposition). If so, these queues are rebuilt by executing S514 through S518, which are analogous to S504 through S508 described above. Flow then proceeds to S520 from S518, or from S512 if the determination therein is negative.
A weighted not-in-use time is determined for the oldest resources of each queue. For example, a weight of 10 may be assigned to resources of a short-term disposition queue, while resources of a long-term disposition queue may be assigned a weight of 1. Moreover, a not-in-use time may be determined for the oldest cached resource in each queue by subtracting the timestamp of the oldest cached resource from the current timestamp. The determined not-in-use time of each oldest cached resource is then multiplied by the weight assigned to its resource disposition to determine its weighted not-in-use time. For example, if the oldest cached “short-term” resource has a not-in-use time of 5 s and the oldest cached “long-term” resource has a not-in-use time of 20 s, the weighted not-in-use time for the short-term resource is 5 s×10=50 s and the weighted not-in-use time for the long-term resource is 20 s×1=20 s.
Next, at S522, the oldest cached resource which is associated with the largest weighted not-in-use time is deallocated. Considering the above example, the short term resource associated with the weighted not-in-use time of 50 s is deallocated at S522. The pointers of the resource's header, as well as any pointers to the resource, remain intact to maintain the integrity of its resource queue. Flow then continues to S524 to determine whether the required amount of resources has been deallocated. If not, flow returns to S512 to determine whether any of the queues are empty and associated with a set resource-present flag and, if not, to deallocate another resource as described above.
If the determination at S524 is positive, flow returns to S510 to wait for another command to deallocate an amount of resources from the cache. Again, the resource queues are not updated or regenerated until all the resources of the queue have been deallocated. Process 500 therefore reduces the number and frequency of locks required to implement a replacement policy, while also providing a policy which may consider the relative priority of different types of resources.
Elements described herein as communicating with one another are directly or indirectly capable of communicating over any number of different systems for transferring data, including but not limited to shared memory communication, a local area network, a wide area network, a telephone network, a cellular network, a fiber-optic network, a satellite network, an infrared network, a radio frequency network, and any other type of network that may be used to transmit information between devices. Moreover, communication between systems may proceed over any one or more transmission protocols that are or become known, such as Asynchronous Transfer Mode (ATM), Internet Protocol (IP), Hypertext Transfer Protocol (HTTP) and Wireless Application Protocol (WAP).
Embodiments described herein are solely for the purpose of illustration. Those in the art will recognize other embodiments may be practiced with modifications and alterations to that described above.

Claims (17)

What is claimed is:
1. A method implemented on a machine having at least one processor for a database system, comprising:
creating a first resource queue indicating an order of resources stored in a memory, the order based on respective timestamps associated with the stored resources;
associating the first resource queue with a first queue timestamp;
receiving a first command to deallocate a first amount of stored resources from the memory;
determining that a first stored resource indicated by the first resource queue is associated with a timestamp earlier than the first queue timestamp;
deallocating the first stored resource from the memory;
receiving a second command to deallocate a second amount of stored resources from the memory;
determining that the first resource queue indicates no stored resources which are associated with a timestamp earlier than the first queue timestamp;
in response to the determination that the first resource queue indicates no stored resources which are associated with a timestamp earlier than the first queue timestamp, creating a second resource queue indicating a second order of second resources stored in the memory, the second order based on respective second timestamps associated with the stored second resources;
creating a third resource queue indicating a third order of resources of a second disposition type stored in the memory, the third order based on respective timestamps associated with the stored resources of the second disposition type; and
associating the third resource queue with a third queue timestamp;
wherein determining that the first stored resource indicated by the first resource queue is associated with a timestamp earlier than the first queue timestamp comprises:
determining a first not-in-use time for a stored resource indicated by the first resource queue based on a first timestamp associated with the stored resource;
determining a first weighted not-in-use time for the stored resource based on the first not-in-use time and on a first weight associated with the first disposition type;
determining a second not-in-use time for a stored resource indicated by the first resource queue based on a first timestamp associated with the stored resource;
determining a second weighted not-in-use time for the stored second resource based on the second not-in-use time and on a second weight associated with the second disposition type; and
determining that the first weighted not-in-use time is greater than the second weighted not-in-use time.
2. A method according to claim 1, further comprising:
prior to receiving the second command, receiving a third command to deallocate a third amount of stored resources from the memory;
determining that a third stored resource indicated by the first resource queue is associated with a timestamp earlier than the first queue timestamp; and
deallocating the third stored resource from the memory.
3. A method according to claim 1, further comprising:
associating the second resource queue with a second queue timestamp;
determining that a second stored resource indicated by the second queue is associated with a timestamp earlier than the second queue timestamp; and
deallocating the second stored resource from the memory.
4. A method according to claim 1, wherein the resource queue comprises a tree structure.
5. A method according to claim 4, wherein a stored resource is associated with a respective resource header comprising a left pointer member to indicate a first child resource of the stored resource within the tree structure and a right pointer member to indicate a second child resource of the stored resource within the tree structure.
6. A method according to claim 1, wherein determining that the first resource queue indicates no stored resources which are associated with a timestamp earlier than the first queue timestamp further comprises:
determining that the memory stores resources of the first disposition type.
7. A non-transitory computer-readable medium storing program code, the program code executable by a computer, the program code comprising:
code to create a first resource queue indicating an order of resources stored in a memory, the order based on respective timestamps associated with the stored resources;
code to associate the first resource queue with a first queue timestamp;
code to receive a first command to deallocate a first amount of stored resources from the memory;
code to determine that a first stored resource indicated by the first resource queue is associated with a timestamp earlier than the first queue timestamp;
code to deallocate the first stored resource from the memory;
code to receive a second command to deallocate a second amount of stored resources from the memory;
code to determine that the first resource queue indicates no stored resources which are associated with a timestamp earlier than the first queue timestamp;
code to create, in response to the determination that the first resource queue indicates no stored resources which are associated with a timestamp earlier than the first queue timestamp, a second resource queue indicating a second order of second resources stored in the memory, the second order based on respective second timestamps associated with the stored second resources;
code to create a third resource queue indicating a third order of resources of a second disposition type stored in the memory, the third order based on respective timestamps associated with the stored resources of the second disposition type; and
code to associate the third resource queue with a third queue timestamp;
wherein the code to determine that the first stored resource indicated by the first resource queue is associated with a timestamp earlier than the first queue timestamp comprises:
code to determine a first not-in-use time for a stored resource indicated by the first resource queue based on a first timestamp associated with the stored resource;
code to determine a first weighted not-in-use time for the stored resource based on the first not-in-use time and on a first weight associated with the first disposition type;
code to determine a second not-in-use time for a stored resource indicated by the first resource queue based on a first timestamp associated with the stored resource;
code to determine a second weighted not-in-use time for the stored second resource based on the second not-in-use time and on a second weight associated with the second disposition type; and
code to determine that the first weighted not-in-use time is greater than the second weighted not-in-use time.
8. A medium according to claim 7, the program code further comprising:
code to receive, prior to receiving the second command, a third command to deallocate a third amount of stored resources from the memory;
code to determine that a third stored resource indicated by the first resource queue is associated with a timestamp earlier than the first queue timestamp; and
code to deallocate the third stored resource from the memory.
9. A medium according to claim 7, the program code further comprising:
code to associate the second resource queue with a second queue timestamp;
code to determine that a second stored resource indicated by the second queue is associated with a timestamp earlier than the second queue timestamp; and
code to deallocate the second stored resource from the memory.
10. A medium according to claim 7, wherein the resource queue comprises a tree structure.
11. A medium according to claim 10, wherein a stored resource is associated with a respective resource header comprising a left pointer member to indicate a first child resource of the stored resource within the tree structure and a right pointer member to indicate a second child resource of the stored resource within the tree structure.
12. A medium according to claim 7, wherein the code to determine that the first resource queue indicates no stored resources which are associated with a timestamp earlier than the first queue timestamp further comprises:
code to determine that the memory stores resources of the first disposition type.
13. A system comprising:
a database cache;
a memory storing processor-executable process steps; and
a processor to execute the processor-executable process steps to cause the system to:
create a first resource queue indicating an order of resources stored in the database cache, the order based on respective timestamps associated with the stored resources;
associate the first resource queue with a first queue timestamp;
receive a first command to deallocate a first amount of stored resources from the database cache;
determine that a first stored resource indicated by the first resource queue is associated with a timestamp earlier than the first queue timestamp;
deallocate the first stored resource from the database cache;
receive a second command to deallocate a second amount of stored resources from the database cache;
determine that the first resource queue indicates no stored resources which are associated with a timestamp earlier than the first queue timestamp;
create, in response to the determination that the first resource queue indicates no stored resources which are associated with a timestamp earlier than the first queue timestamp, a second resource queue indicating a second order of second resources stored in the database cache, the second order based on respective second timestamps associated with the stored second resources;
create a third resource queue indicating a third order of resources of a second disposition type stored in the database cache, the third order based on respective timestamps associated with the stored resources of the second disposition type; and
associate the third resource queue with a third queue timestamp;
wherein the determination that the first stored resource indicated by the first resource queue is associated with a timestamp earlier than the first queue timestamp comprises:
determination of a first not-in-use time for a stored resource indicated by the first resource queue based on a first timestamp associated with the stored resource;
determination of a first weighted not-in-use time for the stored resource based on the first not-in-use time and on a first weight associated with the first disposition type;
determination of a second not-in-use time for a stored resource indicated by the first resource queue based on a first timestamp associated with the stored resource;
determination of a second weighted not-in-use time for the stored second resource based on the second not-in-use time and on a second weight associated with the second disposition type; and
determination that the first weighted not-in-use time is greater than the second weighted not-in-use time.
14. A system according to claim 13, the processor to execute the processor-executable process steps to cause the system to:
receive, prior to receiving the second command, a third command to deallocate a third amount of stored resources from the database cache;
determine that a third stored resource indicated by the first resource queue is associated with a timestamp earlier than the first queue timestamp; and
deallocate the third stored resource from the database cache.
15. A system according to claim 13, the processor to execute the processor-executable process steps to cause the system to:
associate the second resource queue with a second queue timestamp;
determine that a second stored resource indicated by the second queue is associated with a timestamp earlier than the second queue timestamp; and
deallocate the second stored resource from the database cache.
16. A system according to claim 13, wherein the resource queue comprises a tree structure.
17. A system according to claim 16, wherein a stored resource is associated with a respective resource header comprising a left pointer member to indicate a first child resource of the stored resource within the tree structure and a right pointer member to indicate a second child resource of the stored resource within the tree structure.
US13/169,925 2011-06-27 2011-06-27 Replacement policy for resource container Active 2031-11-17 US8572130B2 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
US13/169,925 US8572130B2 (en) 2011-06-27 2011-06-27 Replacement policy for resource container
EP12004744.4A EP2541423B1 (en) 2011-06-27 2012-06-25 Replacement policy for resource container
US14/065,002 US8819074B2 (en) 2011-06-27 2013-10-28 Replacement policy for resource container

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US13/169,925 US8572130B2 (en) 2011-06-27 2011-06-27 Replacement policy for resource container

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US14/065,002 Continuation US8819074B2 (en) 2011-06-27 2013-10-28 Replacement policy for resource container

Publications (2)

Publication Number Publication Date
US20120331019A1 US20120331019A1 (en) 2012-12-27
US8572130B2 true US8572130B2 (en) 2013-10-29

Family

ID=46603477

Family Applications (2)

Application Number Title Priority Date Filing Date
US13/169,925 Active 2031-11-17 US8572130B2 (en) 2011-06-27 2011-06-27 Replacement policy for resource container
US14/065,002 Active US8819074B2 (en) 2011-06-27 2013-10-28 Replacement policy for resource container

Family Applications After (1)

Application Number Title Priority Date Filing Date
US14/065,002 Active US8819074B2 (en) 2011-06-27 2013-10-28 Replacement policy for resource container

Country Status (2)

Country Link
US (2) US8572130B2 (en)
EP (1) EP2541423B1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9558216B2 (en) 2014-11-21 2017-01-31 Sap Se Moving tables across nodes in an in-memory database instance
US10061832B2 (en) * 2016-11-28 2018-08-28 Oracle International Corporation Database tuple-encoding-aware data partitioning in a direct memory access engine

Families Citing this family (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8572130B2 (en) 2011-06-27 2013-10-29 Sap Ag Replacement policy for resource container
US10152423B2 (en) * 2011-10-31 2018-12-11 International Business Machines Corporation Selective population of secondary cache employing heat metrics
US10444998B1 (en) 2013-10-24 2019-10-15 Western Digital Technologies, Inc. Data storage device providing data maintenance services
US9330143B2 (en) * 2013-10-24 2016-05-03 Western Digital Technologies, Inc. Data storage device supporting accelerated database operations
US10230670B1 (en) 2014-11-10 2019-03-12 Google Llc Watermark-based message queue
US9940360B2 (en) * 2015-05-19 2018-04-10 Sybase, Inc. Streaming optimized data processing
US11392320B2 (en) * 2018-07-03 2022-07-19 Western Digital Technologies, Inc. Quality of service based arbitrations optimized for enterprise solid state drives
US10732897B2 (en) 2018-07-03 2020-08-04 Western Digital Technologies, Inc. Quality of service based arbitrations optimized for enterprise solid state drives
CN111427966B (en) * 2020-06-10 2020-09-22 腾讯科技(深圳)有限公司 Database transaction processing method and device and server
US11669498B2 (en) * 2020-07-24 2023-06-06 EMC IP Holding Company LLC Facilitating exclusive local locks on a distributed file system

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5787473A (en) 1995-09-05 1998-07-28 Emc Corporation Cache management system using time stamping for replacement queue
US5974509A (en) * 1996-05-01 1999-10-26 Sun Microsystems, Inc. Method for purging unused data from a cache memory
US20010021964A1 (en) * 1998-09-08 2001-09-13 Jukka-Pekka Iivonen Method for implementing a queue in a memory, and memory arrangement
US6353898B1 (en) * 1997-02-21 2002-03-05 Novell, Inc. Resource management in a clustered computer system
US6397228B1 (en) * 1999-03-31 2002-05-28 Verizon Laboratories Inc. Data enhancement techniques
US6715039B1 (en) 2001-09-12 2004-03-30 Emc Corporation Cache slot promotion in a replacement queue cache using determinations of probabilities and costs
US6834329B2 (en) * 2001-07-10 2004-12-21 Nec Corporation Cache control method and cache apparatus
US20050055512A1 (en) * 2003-09-05 2005-03-10 Kishi Gregory Tad Apparatus, system, and method flushing data from a cache to secondary storage
US20050268044A1 (en) * 2004-06-01 2005-12-01 Arcas Blaise A Y Efficient data cache
US7519776B2 (en) * 2005-07-26 2009-04-14 Invensys Systems, Inc. Method and system for time-weighted cache management
US8117396B1 (en) * 2006-10-10 2012-02-14 Network Appliance, Inc. Multi-level buffer cache management through soft-division of a uniform buffer cache

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8572130B2 (en) 2011-06-27 2013-10-29 Sap Ag Replacement policy for resource container

Patent Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5787473A (en) 1995-09-05 1998-07-28 Emc Corporation Cache management system using time stamping for replacement queue
US5974509A (en) * 1996-05-01 1999-10-26 Sun Microsystems, Inc. Method for purging unused data from a cache memory
US6353898B1 (en) * 1997-02-21 2002-03-05 Novell, Inc. Resource management in a clustered computer system
US20010021964A1 (en) * 1998-09-08 2001-09-13 Jukka-Pekka Iivonen Method for implementing a queue in a memory, and memory arrangement
US6374339B2 (en) * 1998-09-08 2002-04-16 Nokia Networks Oy Method for implementing a queue in a memory, and memory arrangement
US6397228B1 (en) * 1999-03-31 2002-05-28 Verizon Laboratories Inc. Data enhancement techniques
US6834329B2 (en) * 2001-07-10 2004-12-21 Nec Corporation Cache control method and cache apparatus
US6715039B1 (en) 2001-09-12 2004-03-30 Emc Corporation Cache slot promotion in a replacement queue cache using determinations of probabilities and costs
US20050055512A1 (en) * 2003-09-05 2005-03-10 Kishi Gregory Tad Apparatus, system, and method flushing data from a cache to secondary storage
US20050268044A1 (en) * 2004-06-01 2005-12-01 Arcas Blaise A Y Efficient data cache
US7519776B2 (en) * 2005-07-26 2009-04-14 Invensys Systems, Inc. Method and system for time-weighted cache management
US8117396B1 (en) * 2006-10-10 2012-02-14 Network Appliance, Inc. Multi-level buffer cache management through soft-division of a uniform buffer cache

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9558216B2 (en) 2014-11-21 2017-01-31 Sap Se Moving tables across nodes in an in-memory database instance
US10061832B2 (en) * 2016-11-28 2018-08-28 Oracle International Corporation Database tuple-encoding-aware data partitioning in a direct memory access engine

Also Published As

Publication number Publication date
EP2541423B1 (en) 2019-01-09
US20140059082A1 (en) 2014-02-27
EP2541423A1 (en) 2013-01-02
US20120331019A1 (en) 2012-12-27
US8819074B2 (en) 2014-08-26

Similar Documents

Publication Publication Date Title
US8819074B2 (en) Replacement policy for resource container
US11775524B2 (en) Cache for efficient record lookups in an LSM data structure
US10564850B1 (en) Managing known data patterns for deduplication
US9495296B2 (en) Handling memory pressure in an in-database sharded queue
US8868831B2 (en) Caching data between a database server and a storage system
US10275489B1 (en) Binary encoding-based optimizations at datastore accelerators
US9307024B2 (en) Efficient storage of small random changes to data on disk
EP2336901B1 (en) Online access to database snapshots
US20130290636A1 (en) Managing memory
US20120317339A1 (en) System and method for caching data in memory and on disk
WO2011064742A1 (en) Super-records
CN112148736A (en) Method, device and storage medium for caching data
US10146833B1 (en) Write-back techniques at datastore accelerators
US10642745B2 (en) Key invalidation in cache systems
CN117539915B (en) Data processing method and related device
CN111581123B (en) Classification-based locking of memory allocation
EP4168899A1 (en) Sparse file system implemented with multiple cloud services
US8843708B2 (en) Control block linkage for database converter handling
US8549041B2 (en) Converter traversal using power of two-based operations
US9442948B2 (en) Resource-specific control blocks for database cache
KR101368441B1 (en) Apparatus, method and computer readable recording medium for reusing a free space of database
CN106873910B (en) SSD (solid State disk) caching method based on openstack platform
CN116775700A (en) Data caching method, device and storage medium

Legal Events

Date Code Title Description
AS Assignment

Owner name: SAP AG, GERMANY

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SCHRETER, IVAN;REEL/FRAME:026507/0888

Effective date: 20110622

FEPP Fee payment procedure

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCF Information on status: patent grant

Free format text: PATENTED CASE

AS Assignment

Owner name: SAP SE, GERMANY

Free format text: CHANGE OF NAME;ASSIGNOR:SAP AG;REEL/FRAME:033625/0334

Effective date: 20140707

FEPP Fee payment procedure

Free format text: PAYER NUMBER DE-ASSIGNED (ORIGINAL EVENT CODE: RMPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

FPAY Fee payment

Year of fee payment: 4

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 8