US20080097997A1 - System and method for optimistic caching - Google Patents

System and method for optimistic caching Download PDF

Info

Publication number
US20080097997A1
US20080097997A1 US11/955,229 US95522907A US2008097997A1 US 20080097997 A1 US20080097997 A1 US 20080097997A1 US 95522907 A US95522907 A US 95522907A US 2008097997 A1 US2008097997 A1 US 2008097997A1
Authority
US
United States
Prior art keywords
data item
database
transaction
data
computer implemented
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US11/955,229
Inventor
Seth White
Adam Messinger
Dean Jacobs
Rob Woollen
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Oracle International Corp
Original Assignee
BEA Systems Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by BEA Systems Inc filed Critical BEA Systems Inc
Priority to US11/955,229 priority Critical patent/US20080097997A1/en
Publication of US20080097997A1 publication Critical patent/US20080097997A1/en
Assigned to ORACLE INTERNATIONAL CORPORATION reassignment ORACLE INTERNATIONAL CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: BEA SYSTEMS, INC.
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/08Addressing or allocation; Relocation in hierarchically structured memory systems, e.g. virtual memory systems
    • G06F12/0802Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches
    • G06F12/0806Multiuser, multiprocessor or multiprocessing cache systems
    • G06F12/0815Cache consistency protocols
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S707/00Data processing: database and file management or data structures
    • Y10S707/99931Database or file accessing
    • Y10S707/99938Concurrency, e.g. lock management in shared database
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S707/00Data processing: database and file management or data structures
    • Y10S707/99951File or database maintenance
    • Y10S707/99952Coherency, e.g. same view to multiple users

Definitions

  • the invention relates to the caching of data for multiple transactions.
  • a concurrency model such as exclusive concurrency can be used to provide such access.
  • each server instance manages an exclusive lock on a primary key and any data associated with that key. This lock is held for the length of the transaction when invoked for a transaction.
  • a non-transactional invoke such as for a method that requires access to, but will not update, the data, the lock is held until the method completes.
  • concurrency control is typically provided by the database.
  • the exclusive concurrency model works well in a single server system in which the server has exclusive access to the database.
  • a sharing value such as “db-is-shared” can be set to false, such that the any read of the data can read from an instance in an in-memory cache instead of reading from the database.
  • Deadlocking can occur when users have locks on separate objects, and one of the users is trying to acquire a lock on an object that is locked by another user. Beans holding instances of the data can also deadlock themselves if the beans are called within a transaction and subsequently called without a transaction.
  • a first instance of a data item is read into a first cache for a first transaction.
  • the first data item can be read from the database or from the cache for a previous transaction.
  • a second instance of the data item is read into a second cache for a second transaction.
  • the second cache can read from the first cache, from the database, or from a cache for a previous transaction.
  • the data item in the database can then be updated by committing changes for one of the transactions. The changes can also be committed to the cached instance of the data item for that transaction.
  • An optimistic concurrency algorithm directs the updating cache to notify the other cache that the data item has changed and that the other instance is not current.
  • the outdated cache can then drop its instance of the data item and read a new instance.
  • the outdated cache can read the new instance from the database, from the updated cache, or from a cache for a previous transaction.
  • FIG. 1 is a diagram of a system in accordance with one embodiment of the present invention.
  • FIG. 2 is a diagram of a system in accordance with another embodiment of the present invention.
  • FIG. 3 is a flowchart showing the steps of a method in accordance with the embodiment of the FIG. 1 .
  • Systems and methods in accordance with one embodiment of the present invention can overcome deficiencies in prior art concurrency models by utilizing “optimistic caching” or “optimistic concurrency”. Using optimistic concurrency, it is not necessary for each transaction to read from the database and consume system resources unnecessarily.
  • the approach is referred to herein as “optimistic” because a server instance does not get a lock on the data being used by a transaction. Transactions are optimistic about the fact that no changes are going to be made to the data while it is being used by that transaction. Under this belief, there is no need to lock the data since it will not change during the transaction. Since the data is not locked, it can be used concurrently by multiple users.
  • An invalidation scheme can be used to facilitate optimistic caching in a clustered environment, such that multiple machines in a cluster can run an optimistic algorithm in parallel. When changes are made, the machines can communicate with each other that a change has occurred, and can indicate what those changes are. This can help to avoid an optimistic conflict or exception at the end of a transaction.
  • XML can be used with such a system, as XML provides a declarative way for a user to indicate that optimistic caching is to be used.
  • FIG. 1 Such a system is shown in FIG. 1 .
  • a data item 102 is stored in a database 102 .
  • User 106 and user 110 may both want access to the data item 102 concurrently. Since the data item is not locked, an instance of the data item can be read into a cache or bean instance 104 for user 106 , and a second instance can be read into a cache or bean instance 108 for user 110 .
  • the optimistic algorithm can direct cache 104 to notify cache 108 .
  • the algorithm can also instruct the transaction for user 106 to update the instance of the data item in cache 104 . If the instance in cache 104 is updated, cache 108 can read a new instance from cache 104 . If cache 104 is not updated, cache 110 can read from the database 100 .
  • nodes in a cluster there can be multiple nodes in a cluster that are able to access and make changes to the data in the database, as shown in FIG. 2 .
  • an application can evenly utilize any available cluster nodes 204 , 206 , 208 .
  • a user 200 that wishes to utilize data in the database 202 can be directed to work through node 204 .
  • the user is not aware of which node is being utilized, as the nodes appear to the user as a single server.
  • An object can migrate from one node to another in support of even load distribution.
  • node 204 is shown to be able to send an update message to nodes 206 and 208 in the event that the user 200 updates data in the database 202 .
  • Such information can prevent other nodes from having to wait until the end of a transaction to find out that the data has changed. If a node does not find out about the change until the end of a transaction, the node will need to rollback the transaction, obtain the updated state of the data, and restart the processing of the transaction.
  • This notification of nodes in a cluster can save both time and resources by preventing the reprocessing of transactions. Different schemes can be used that allow the nodes to notify themselves of changes in the data.
  • FIG. 3 shows steps for a method that could be used with a system in accordance with the embodiment of FIG. 1 .
  • an instance of a data item is read into a first cache for a first transaction 300 .
  • An instance of the data item is then read from the first cache into a second cache for a second transaction 302 .
  • One of the transactions can update the data item in the database by committing changes to the data item, and can also update the associated instance in cache 304 .
  • a notification can then be sent to the cache for the other transaction notifying the cache that the data item has changed 306 . That cache can then drop its instance of the data item and read a new instance from the database or from the updated cache 308 .
  • Certain operations can be done outside of a transaction.
  • One such operation involves suspending the active transaction while reading data.
  • This operation can be a global transaction, for example, that spans multiple nodes about the network and can be committed using a two-phase commit algorithm.
  • “Creates” and “removes” can be done within the context of global transactions, such that if a global transaction rolls back then the operations can be rolled back as well. Long-term locks can be held during these operations.
  • a copy of the data can be made so that the data is read during a short-term transaction. An initial copy of the data being read can be made. The transaction can then actually update a different copy of the data, such that at the end of the transaction those copies can be compared with what is in the database, in order to help determine whether to commit the transaction.
  • This optimistic caching can also be done with Enterprise JavaBeans (EJBs).
  • EJBs Enterprise JavaBeans
  • An EJB has a life cycle, and certain embodiments of optimistic caching work within the context of this life cycle.
  • the EJBs can load themselves from the database and can store themselves in the database.
  • the EJBs can also undergo other operations, such as loading, storing, creating, removing, passivating, and activating.
  • EJBs can also read data in a local transaction. This allows the EJBs to read data items without acquiring long-term locks on the data.
  • optimistic caching allows for the caching of data between separate transactions.
  • entity beans holding instances of the data do not perform any better than stateless session beans.
  • caching can be done between transactions to the extent possible.
  • One way to accomplish between-transaction caching is to read data in a local transaction so there is no need for long-term locks. At the end of the transaction, the data can be checked to determine if anyone has made any changes.
  • a “db-is-shared” or equivalent tag can be replaced with a tag such as “AppServerCachingEnabled”, which can allow the caching of entity beans between transactions when AppServerCachingEnabled has a value of “true”.
  • each transaction can activate its own bean instance. There is no locking within the application server.
  • Two new tags can be used, such as “OptimisticVerifyModifiedFields” and “OptimisticVerifyReadFields”.
  • OptimisticVerifyReadFields the update statement can verify that every field that was read in the transaction matches the current content in the database.
  • OptimisticVerifyModifiedFields can be used to check only the fields that have been updated. The transaction can abort if the update conditions fail. If the transaction is a read-only transaction, it will not produce an update statement and there will be no verification.
  • Such a system can utilize special abstract methods for CMP fields such as “increment”, “decrement”, and “decrementGreaterThan”. For example, if there is an item being purchased at an online store by several users concurrently, each individual transaction simply wants to decrease the inventory amount by one unit. Each transaction is able to commit the update as long as the inventory is greater than zero units. The transaction does not care about how many units are in inventory, as long as there is at least one. Therefore, an operator could be used such as:
  • a SQL statement could then be issued such as:
  • OptimisticVerifyModifiedFields and OptimisticVerifyReadFields could also be used as the basis for clustered entity bean caching.
  • a server could keep a cache that associates a primary key and field group with cached data.
  • a field group can be a set of container-managed fields. When one member of a group needs to be read from the database, the entire group is read. By default, the CMP fields can be in a single group, but this can be changed using a deployment descriptor, for example.
  • an entity bean When an entity bean reads a field group, it can ask the cache for the field group. If the cache has the field group in memory, it can return the cached data instead of fetching the data from the database. When an updating transaction commits, the changes can be entered into the cache as well as into the database. This allows the cache to have the most recent data without having to hit the database.
  • the cache can also measure the hit rate and optimistic rollback rate. These statistics can be used to drop or pre-fetch cached data.
  • the message can contain, for example, the JNDI name, the primary key, and the field group number.
  • the cache can drop the corresponding field group from its cache. If this field group has a high hit rate, the cache can pre-fetch the new version.
  • the predicated update can be used to ensure that the read or modified fields are consistent with that which is still in the database.

Abstract

Transactions are granted concurrent access to a data item through the use of an optimistic concurrency algorithm. Each transaction gets its own instance of the data item, such as in a cache or in an entity bean, such that it is not necessary to lock the data. The instances can come from the data or from other instances. When a transaction updates the data item, the optimistic concurrency algorithm ensures that the other instances are notified that the data item has been changed and that it is necessary to read a new instance, from the database or from an updated instance.

Description

    CLAIM OF PRIORITY
  • This application is a continuation of U.S. patent application Ser. No. 11/339,180, entitled “SYSTEM AND METHOD FOR OPTIMISTIC CACHING” by White et al., filed Jan. 25, 2006; which is a continuation of U.S. patent application Ser. No. 10/340,023, now U.S. Pat. No. 7,020,684 entitled “SYSTEM AND METHOD FOR OPTIMISTIC CACHING” by White et al., filed Jan. 10, 2003; which claims priority from U.S. Provisional Patent Application No. 60/349,464 entitled “SYSTEM AND METHOD FOR OPTIMISTIC CACHING” by White, et al. filed Jan. 18, 2002, incorporated herein by reference.
  • COPYRIGHT NOTICE
  • A portion of the disclosure of this patent document contains material which is subject to copyright protection. The copyright owner has no objection to the facsimile reproduction by anyone of the patent document of the patent disclosure, as it appears in the Patent and Trademark Office patent file or records, but otherwise reserves all copyright rights whatsoever.
  • CROSS REFERENCE TO RELATED PATENT DOCUMENTS
  • The following co-pending U.S. patent documents are assigned to BEA Systems, Inc., the assignee of the present application, and these documents are hereby incorporated herein by reference. U.S. patent application Ser. No. 10/340,067 filed Jan. 10, 2003, now U.S. Pat. No. 6,978,278 to Seth White et al. and entitled, “System and Method for Heterogeneous Caching” and U.S. patent application Ser. No. 10/340,301 filed Jan. 10, 2003, now U.S. Pat. No. 7,136,879 to Seth White et al. and entitled, “System and Method for Read-Only Entity Bean Caching”; and U.S. patent application Ser. No. 09/975,590 filed Oct. 11, 2001, to Dean Jacobs et al. and entitled “Data Replication Protocol”.
  • FIELD OF THE INVENTION
  • The invention relates to the caching of data for multiple transactions.
  • BACKGROUND
  • In a system where multiple users desire concurrent access to the same data, a concurrency model such as exclusive concurrency can be used to provide such access. In such a model, each server instance manages an exclusive lock on a primary key and any data associated with that key. This lock is held for the length of the transaction when invoked for a transaction. In the case of a non-transactional invoke, such as for a method that requires access to, but will not update, the data, the lock is held until the method completes. In a clustered environment, it is possible that a given primary key is active in multiple servers. In such a situation, concurrency control is typically provided by the database.
  • The exclusive concurrency model works well in a single server system in which the server has exclusive access to the database. A sharing value, such as “db-is-shared” can be set to false, such that the any read of the data can read from an instance in an in-memory cache instead of reading from the database.
  • Problems arise; however, as the exclusive concurrency model requires the use of deadlocks. Deadlocking can occur when users have locks on separate objects, and one of the users is trying to acquire a lock on an object that is locked by another user. Beans holding instances of the data can also deadlock themselves if the beans are called within a transaction and subsequently called without a transaction.
  • Other systems provide the ability for each transaction to activate a unique instance of the data. These systems do not provide concurrency control within the application server, however, but rely on the database for concurrency. This database concurrency model is an improvement over exclusive concurrency, but it is still necessary to hit the database for every transaction. This approach can still strain system resources under load.
  • BRIEF SUMMARY
  • Systems and methods in accordance with embodiments of the present invention can overcome these and other deficiencies in the prior art by providing an optimistic approach to concurrency. In one embodiment, a first instance of a data item is read into a first cache for a first transaction. The first data item can be read from the database or from the cache for a previous transaction. A second instance of the data item is read into a second cache for a second transaction. The second cache can read from the first cache, from the database, or from a cache for a previous transaction. The data item in the database can then be updated by committing changes for one of the transactions. The changes can also be committed to the cached instance of the data item for that transaction. An optimistic concurrency algorithm directs the updating cache to notify the other cache that the data item has changed and that the other instance is not current. The outdated cache can then drop its instance of the data item and read a new instance. The outdated cache can read the new instance from the database, from the updated cache, or from a cache for a previous transaction.
  • Other features, aspects and objects of the invention can be obtained from a review of the specification, the figures, and the claims.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is a diagram of a system in accordance with one embodiment of the present invention.
  • FIG. 2 is a diagram of a system in accordance with another embodiment of the present invention.
  • FIG. 3 is a flowchart showing the steps of a method in accordance with the embodiment of the FIG. 1.
  • DETAILED DESCRIPTION
  • Systems and methods in accordance with one embodiment of the present invention can overcome deficiencies in prior art concurrency models by utilizing “optimistic caching” or “optimistic concurrency”. Using optimistic concurrency, it is not necessary for each transaction to read from the database and consume system resources unnecessarily.
  • The approach is referred to herein as “optimistic” because a server instance does not get a lock on the data being used by a transaction. Transactions are optimistic about the fact that no changes are going to be made to the data while it is being used by that transaction. Under this belief, there is no need to lock the data since it will not change during the transaction. Since the data is not locked, it can be used concurrently by multiple users.
  • In order to assure the accuracy of transactions using that data, however, it will be necessary to determine whether the underlying data actually changed before transactional changes are written to the database.
  • An invalidation scheme can be used to facilitate optimistic caching in a clustered environment, such that multiple machines in a cluster can run an optimistic algorithm in parallel. When changes are made, the machines can communicate with each other that a change has occurred, and can indicate what those changes are. This can help to avoid an optimistic conflict or exception at the end of a transaction. XML can be used with such a system, as XML provides a declarative way for a user to indicate that optimistic caching is to be used.
  • Such a system is shown in FIG. 1. A data item 102 is stored in a database 102. User 106 and user 110 may both want access to the data item 102 concurrently. Since the data item is not locked, an instance of the data item can be read into a cache or bean instance 104 for user 106, and a second instance can be read into a cache or bean instance 108 for user 110. If user 106 updates the data item 102, the optimistic algorithm can direct cache 104 to notify cache 108. The algorithm can also instruct the transaction for user 106 to update the instance of the data item in cache 104. If the instance in cache 104 is updated, cache 108 can read a new instance from cache 104. If cache 104 is not updated, cache 110 can read from the database 100.
  • There can be multiple nodes in a cluster that are able to access and make changes to the data in the database, as shown in FIG. 2. In order to make efficient use of a cluster, an application can evenly utilize any available cluster nodes 204, 206, 208. For instance, a user 200 that wishes to utilize data in the database 202 can be directed to work through node 204. The user is not aware of which node is being utilized, as the nodes appear to the user as a single server. An object can migrate from one node to another in support of even load distribution.
  • It can be beneficial for these nodes to notify each other about any changes made to data. In FIG. 2, node 204 is shown to be able to send an update message to nodes 206 and 208 in the event that the user 200 updates data in the database 202. Such information can prevent other nodes from having to wait until the end of a transaction to find out that the data has changed. If a node does not find out about the change until the end of a transaction, the node will need to rollback the transaction, obtain the updated state of the data, and restart the processing of the transaction. This notification of nodes in a cluster can save both time and resources by preventing the reprocessing of transactions. Different schemes can be used that allow the nodes to notify themselves of changes in the data.
  • FIG. 3 shows steps for a method that could be used with a system in accordance with the embodiment of FIG. 1. In this method, an instance of a data item is read into a first cache for a first transaction 300. An instance of the data item is then read from the first cache into a second cache for a second transaction 302. One of the transactions can update the data item in the database by committing changes to the data item, and can also update the associated instance in cache 304. A notification can then be sent to the cache for the other transaction notifying the cache that the data item has changed 306. That cache can then drop its instance of the data item and read a new instance from the database or from the updated cache 308.
  • Certain operations can be done outside of a transaction. One such operation involves suspending the active transaction while reading data. This operation can be a global transaction, for example, that spans multiple nodes about the network and can be committed using a two-phase commit algorithm.
  • “Creates” and “removes” can be done within the context of global transactions, such that if a global transaction rolls back then the operations can be rolled back as well. Long-term locks can be held during these operations. During the load, a copy of the data can be made so that the data is read during a short-term transaction. An initial copy of the data being read can be made. The transaction can then actually update a different copy of the data, such that at the end of the transaction those copies can be compared with what is in the database, in order to help determine whether to commit the transaction.
  • Enterprise JavaBeans
  • This optimistic caching can also be done with Enterprise JavaBeans (EJBs). An EJB has a life cycle, and certain embodiments of optimistic caching work within the context of this life cycle. The EJBs can load themselves from the database and can store themselves in the database. The EJBs can also undergo other operations, such as loading, storing, creating, removing, passivating, and activating. EJBs can also read data in a local transaction. This allows the EJBs to read data items without acquiring long-term locks on the data.
  • Caching Between Transactions
  • In one embodiment, optimistic caching allows for the caching of data between separate transactions. In prior art systems where every transaction reads from the database, entity beans holding instances of the data do not perform any better than stateless session beans. In order to better utilize the functionality and advantages of entity beans, caching can be done between transactions to the extent possible.
  • One way to accomplish between-transaction caching is to read data in a local transaction so there is no need for long-term locks. At the end of the transaction, the data can be checked to determine if anyone has made any changes. In one such system, a “db-is-shared” or equivalent tag can be replaced with a tag such as “AppServerCachingEnabled”, which can allow the caching of entity beans between transactions when AppServerCachingEnabled has a value of “true”.
  • In a more optimistic embodiment for caching between transactions, data is not read at the beginning of a transaction. It is assumed that what is in cache from a previous transaction is still current with what is in the database. The currency of the data is not checked until the end of the transaction. This allows the transaction to use what is already in cache memory, which can greatly increase the speed of the transaction. If the transaction attempts to commit and finds the data has changed, it can update the data in cache so that the next transaction can read from cache and have the appropriate value.
  • CMP Optimistic Concurrency without Caching
  • There are other optimistic concurrency embodiments that can support the EJB 2.0 Container-Managed Persistence (CMP) specification, and can be used with or without caching between transactions.
  • In an optimistic concurrency approach without caching, each transaction can activate its own bean instance. There is no locking within the application server. Two new tags can be used, such as “OptimisticVerifyModifiedFields” and “OptimisticVerifyReadFields”. With these options, a database read can occur as a separate local transaction. In the case of OptimisticVerifyReadFields, the update statement can verify that every field that was read in the transaction matches the current content in the database. OptimisticVerifyModifiedFields can be used to check only the fields that have been updated. The transaction can abort if the update conditions fail. If the transaction is a read-only transaction, it will not produce an update statement and there will be no verification.
  • Such a system can utilize special abstract methods for CMP fields such as “increment”, “decrement”, and “decrementGreaterThan”. For example, if there is an item being purchased at an online store by several users concurrently, each individual transaction simply wants to decrease the inventory amount by one unit. Each transaction is able to commit the update as long as the inventory is greater than zero units. The transaction does not care about how many units are in inventory, as long as there is at least one. Therefore, an operator could be used such as:
  • public abstract void decrementInventoryIfGreaterThan(int n);
  • A SQL statement could then be issued such as:
  • update set inventory=inventory−1 where inventory>0 and pk=‘item’;
  • CMP Optimistic Concurrency with Caching
  • The OptimisticVerifyModifiedFields and OptimisticVerifyReadFields could also be used as the basis for clustered entity bean caching. A server could keep a cache that associates a primary key and field group with cached data. In a CMP 2.0 implementation, a field group can be a set of container-managed fields. When one member of a group needs to be read from the database, the entire group is read. By default, the CMP fields can be in a single group, but this can be changed using a deployment descriptor, for example.
  • When an entity bean reads a field group, it can ask the cache for the field group. If the cache has the field group in memory, it can return the cached data instead of fetching the data from the database. When an updating transaction commits, the changes can be entered into the cache as well as into the database. This allows the cache to have the most recent data without having to hit the database. The cache can also measure the hit rate and optimistic rollback rate. These statistics can be used to drop or pre-fetch cached data.
  • Multicast Invalidation
  • It can be useful to prevent some of the optimistic rollbacks that can occur if, for example, a first server updates a value and then a second server reads from the second server's outdated or “stale” cache. One way to prevent this is to send invalidation messages between the caches. When an update commits, an invalidation can be sent to the other caches, such as by multicast. The message can contain, for example, the JNDI name, the primary key, and the field group number.
  • When an ejb cache receives an update message, the cache can drop the corresponding field group from its cache. If this field group has a high hit rate, the cache can pre-fetch the new version.
  • While these invalidation messages can help prevent optimistic rollbacks, they do not ensure correctness. The predicated update can be used to ensure that the read or modified fields are consistent with that which is still in the database.
  • The foregoing description of preferred embodiments of the present invention has been provided for the purposes of illustration and description. It is not intended to be exhaustive or to limit the invention to the precise forms disclosed. Many modifications and variations will be apparent to one of ordinary skill in the relevant arts. The embodiments were chosen and described in order to best explain the principles of the invention and its practical application, thereby enabling others skilled in the art to understand the invention for various embodiments and with various modifications that are suited to the particular use contemplated. It is intended that the scope of the invention be defined by the claims and their equivalence.

Claims (20)

1. A computer implemented system for concurrently caching instances of a data item, comprising:
a database capable of storing a data item;
a plurality of cluster nodes capable of executing a transaction using the data item;
a cache for cluster nodes executing a transaction, the cache capable of storing an instance of the data item; and
an optimistic caching algorithm running concurrently on multiple cluster nodes, the algorithm capable of directing the clusters node to notify other cluster nodes when updating the data item in the database.
2. The computer implemented system according to claim 1, wherein:
the cluster nodes are further capable of checking the data item in the database before updating the data item.
3. The computer implemented system according to claim 1, wherein:
the instances of data items are stored as enterprise beans at the caches.
4. The computer implemented system according to claim 3, wherein:
the instances of data items are stored as enterprise java beans at the caches.
5. The computer implemented system according to claim 4, wherein:
the enterprise java beans load themselves from the database.
6. The computer implemented system according to claim 4, wherein:
the enterprise java beans store themselves in the database.
7. The computer implemented system according to claim 1, wherein:
the currency of the data is checked when a transaction commits.
8. A computer implemented system for concurrently caching instances of a data item, comprising:
a plurality of nodes capable of executing a transaction wherein the transaction is adapted to use an instance of a data item derived from a data item stored in a database;
at least one cache associated with the plurality of nodes being capable of storing the instance of a data item; and
an optimistic caching algorithm running concurrently on at least some of the plurality of nodes, the optimistic caching algorithm capable of directing nodes to notify other nodes when updating the data item in the database.
9. The system of claim 8, further comprising:
a plurality of caches, wherein each cache is associated with one of the plurality of nodes.
10. The system of claim 8, wherein:
the instances of data items in the plurality of caches can be updated from at least one of the data item stored in the database and the instance of the data item associated with another of said plurality of caches.
11. The computer implemented system according to claim 8, wherein:
the cluster nodes are further capable of checking the data item in the database before updating the data item.
12. The computer implemented system according to claim 8, wherein:
the instances of data items are stored as enterprise beans at the caches.
13. The computer implemented system according to claim 12, wherein:
the instances of data items are stored as enterprise java beans at the caches.
14. The computer implemented system according to claim 13, wherein:
the enterprise java beans load themselves from the database.
15. The computer implemented system according to claim 13, wherein:
the enterprise java beans store themselves in the database.
16. The computer implemented system according to claim 8, wherein:
the currency of the data is checked when a transaction commits.
17. A computer implemented system for concurrently caching instances of a data item, comprising:
a plurality of nodes capable of executing a transaction wherein the transaction is adapted to use an instance of a data item derived from a data item stored in a database;
at least one cache associated with the plurality of nodes being capable of storing the instance of a data item; and
an optimistic caching algorithm running concurrently on at least some of the plurality of nodes, the optimistic caching algorithm capable of directing nodes to notify other nodes when updating the data item in the database, data items are stored as enterprise beans at the at least one cache.
18. The computer implemented system according to claim 17, wherein:
the instances of data items are stored as enterprise java beans at the caches.
19. The computer implemented system according to claim 18, wherein:
the enterprise java beans load themselves from the database.
20. The computer implemented system according to claim 18, wherein:
the enterprise java beans store themselves in the database.
US11/955,229 2002-01-18 2007-12-12 System and method for optimistic caching Abandoned US20080097997A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US11/955,229 US20080097997A1 (en) 2002-01-18 2007-12-12 System and method for optimistic caching

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US34946402P 2002-01-18 2002-01-18
US10/340,023 US7020684B2 (en) 2002-01-18 2003-01-10 System and method for optimistic caching
US11/339,180 US7328322B2 (en) 2002-01-18 2006-01-25 System and method for optimistic caching
US11/955,229 US20080097997A1 (en) 2002-01-18 2007-12-12 System and method for optimistic caching

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US11/339,180 Continuation US7328322B2 (en) 2002-01-18 2006-01-25 System and method for optimistic caching

Publications (1)

Publication Number Publication Date
US20080097997A1 true US20080097997A1 (en) 2008-04-24

Family

ID=27616675

Family Applications (3)

Application Number Title Priority Date Filing Date
US10/340,023 Expired - Lifetime US7020684B2 (en) 2002-01-18 2003-01-10 System and method for optimistic caching
US11/339,180 Expired - Lifetime US7328322B2 (en) 2002-01-18 2006-01-25 System and method for optimistic caching
US11/955,229 Abandoned US20080097997A1 (en) 2002-01-18 2007-12-12 System and method for optimistic caching

Family Applications Before (2)

Application Number Title Priority Date Filing Date
US10/340,023 Expired - Lifetime US7020684B2 (en) 2002-01-18 2003-01-10 System and method for optimistic caching
US11/339,180 Expired - Lifetime US7328322B2 (en) 2002-01-18 2006-01-25 System and method for optimistic caching

Country Status (2)

Country Link
US (3) US7020684B2 (en)
WO (1) WO2003063026A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100161571A1 (en) * 2008-12-18 2010-06-24 Winfried Schwarzmann Ultimate locking mechanism

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6978278B2 (en) * 2002-01-18 2005-12-20 Bea Systems, Inc. System and method for heterogeneous caching
US8655755B2 (en) 2003-10-22 2014-02-18 Scottrade, Inc. System and method for the automated brokerage of financial instruments
US7454571B1 (en) 2004-05-04 2008-11-18 Sun Microsystems, Inc. Heuristic cache tuning
US20060004688A1 (en) * 2004-07-02 2006-01-05 Scanlon James R Systems and methods for integrating various processes
US7502843B2 (en) * 2004-12-30 2009-03-10 Microsoft Corporation Server queuing system and method
JP5772458B2 (en) * 2011-09-29 2015-09-02 富士通株式会社 Data management program, node, and distributed database system
US10284649B2 (en) * 2013-05-31 2019-05-07 Nec Corporation Distributed processing system
US9117189B2 (en) * 2013-08-23 2015-08-25 Oracle International Corporation System and method for object lock management using cached lock objects
CN104657483B (en) * 2015-02-28 2018-06-15 华为技术有限公司 Handle method, processing node, Centroid and the cluster of affairs
US10313256B2 (en) * 2015-05-21 2019-06-04 Intel Corporation Apparatus and methods for adaptive data compression
US10282115B2 (en) 2017-04-13 2019-05-07 International Business Machines Corporation Object synchronization in a clustered system

Citations (54)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5261069A (en) * 1990-08-13 1993-11-09 Hewlett-Packard Company Method of maintaining consistency of cached data in a database system
US5581753A (en) * 1994-09-28 1996-12-03 Xerox Corporation Method for providing session consistency guarantees
US5613060A (en) * 1990-05-16 1997-03-18 International Business Machines Corporation Asynchronous resynchronization of a commit procedure
US5634052A (en) * 1994-10-24 1997-05-27 International Business Machines Corporation System for reducing storage requirements and transmission loads in a backup subsystem in client-server environment by transmitting only delta files from client to server
US5765171A (en) * 1995-12-29 1998-06-09 Lucent Technologies Inc. Maintaining consistency of database replicas
US5813017A (en) * 1994-10-24 1998-09-22 International Business Machines Corporation System and method for reducing storage requirement in backup subsystems utilizing segmented compression and differencing
US5878213A (en) * 1996-02-15 1999-03-02 International Business Machines Corporation Methods, systems and computer program products for the synchronization of time coherent caching system
US5909689A (en) * 1997-09-18 1999-06-01 Sony Corporation Automatic update of file versions for files shared by several computers which record in respective file directories temporal information for indicating when the files have been created
US6012059A (en) * 1997-08-21 2000-01-04 Dataxel Corporation Method and apparatus for replicated transaction consistency
US6065046A (en) * 1997-07-29 2000-05-16 Catharon Productions, Inc. Computerized system and associated method of optimally controlled storage and transfer of computer programs on a computer network
US6086632A (en) * 1996-10-31 2000-07-11 Nec Corporation Register optimizing compiler using commutative operations
US6240413B1 (en) * 1997-12-22 2001-05-29 Sun Microsystems, Inc. Fine-grained consistency mechanism for optimistic concurrency control using lock groups
US6256634B1 (en) * 1998-06-30 2001-07-03 Microsoft Corporation Method and system for purging tombstones for deleted data items in a replicated database
US6266742B1 (en) * 1997-10-27 2001-07-24 International Business Machines Corporation Algorithm for cache replacement
US6304879B1 (en) * 1998-11-25 2001-10-16 Microsoft Corporation Dynamic data cache for object-oriented computing environments
US20010042073A1 (en) * 1999-06-22 2001-11-15 Christian D. Saether Method and system for automatically updating the version of a set of files stored on content servers
US20020004850A1 (en) * 2000-03-29 2002-01-10 Krishna Sudarshan System and method of providing a messaging engine for an enterprise javabeans enabled server to achieve container managed asynchronous functionality
US6366930B1 (en) * 1996-04-12 2002-04-02 Computer Associates Think, Inc. Intelligent data inventory & asset management systems method and apparatus
US6401239B1 (en) * 1999-03-22 2002-06-04 B.I.S. Advanced Software Systems Ltd. System and method for quick downloading of electronic files
US20020073188A1 (en) * 2000-12-07 2002-06-13 Rawson Freeman Leigh Method and apparatus for partitioning system management information for a server farm among a plurality of leaseholds
US6430564B1 (en) * 1999-03-01 2002-08-06 Hewlett-Packard Company Java data manager for embedded device
US20020107934A1 (en) * 2001-01-12 2002-08-08 Epicrealm Inc. Method and system for dynamic distributed data caching
US6453321B1 (en) * 1999-02-11 2002-09-17 Ibm Corporation Structured cache for persistent objects
US20020184444A1 (en) * 2000-12-22 2002-12-05 Shandony Michael J. Request based caching of data store data
US20020188591A1 (en) * 2001-06-08 2002-12-12 International Business Machines Corporation Disabling and reloading enterprise java beans using database trigger programs
US20030014480A1 (en) * 2001-07-16 2003-01-16 Sam Pullara Method and apparatus for session replication and failover
US6526521B1 (en) * 1999-06-18 2003-02-25 Emc Corporation Methods and apparatus for providing data storage access
US20030050972A1 (en) * 2001-07-17 2003-03-13 Felt Edward P. System and method for transaction processing with transaction property feature
US6542926B2 (en) * 1998-06-10 2003-04-01 Compaq Information Technologies Group, L.P. Software partitioned multi-processor system with flexible resource sharing levels
US20030065826A1 (en) * 2001-09-06 2003-04-03 Jim Skufca System and method for dynamically caching dynamic multi-sourced persisted EJBs
US20030074580A1 (en) * 2001-03-21 2003-04-17 Knouse Charles W. Access system interface
US6578160B1 (en) * 2000-05-26 2003-06-10 Emc Corp Hopkinton Fault tolerant, low latency system resource with high level logging of system resource transactions and cross-server mirrored high level logging of system resource transactions
US20030110467A1 (en) * 2001-04-20 2003-06-12 Sree Ayyanar Spinning And Weaving Mills Limited Data storage schema independent programming for data retrieval using semantic bridge
US20030115366A1 (en) * 2001-12-18 2003-06-19 Robinson Brian R. Asynchronous message delivery system and method
US6609213B1 (en) * 2000-08-10 2003-08-19 Dell Products, L.P. Cluster-based system and method of recovery from server failures
US20030200350A1 (en) * 2002-04-19 2003-10-23 Ajay Kumar Class dependency graph-based class loading and reloading
US6651140B1 (en) * 2000-09-01 2003-11-18 Sun Microsystems, Inc. Caching pattern and method for caching in an object-oriented programming environment
US6757708B1 (en) * 2000-03-03 2004-06-29 International Business Machines Corporation Caching dynamic content
US20040230747A1 (en) * 2000-07-06 2004-11-18 Ims Steven D. Object caching and update queuing technique to improve performance and resource utilization
US6823514B1 (en) * 2000-11-14 2004-11-23 International Business Machines Corporation Method and system for caching across multiple contexts
US6826601B2 (en) * 2001-09-06 2004-11-30 Bea Systems, Inc. Exactly one cache framework
US6836889B1 (en) * 1999-08-20 2004-12-28 International Business Machines Corporation Code wrapping to simplify access to and use of enterprise JAVA beans
US6898587B2 (en) * 2002-01-18 2005-05-24 Bea Systems, Inc. System and method for performing commutative operations in data access systems
US6918013B2 (en) * 2001-07-16 2005-07-12 Bea Systems, Inc. System and method for flushing bean cache
US6990526B1 (en) * 2000-05-22 2006-01-24 Pointred Technologies, Inc. Method and apparatus for web caching
US7000019B2 (en) * 2000-10-17 2006-02-14 Hewlett-Packard/Development Company L.P. Establishment of a deferred network communication session
US7028030B2 (en) * 2001-08-30 2006-04-11 Bea Systems, Inc. Cluster caching with concurrency checking
US7065616B2 (en) * 2001-02-13 2006-06-20 Network Appliance, Inc. System and method for policy based storage provisioning and management
US20060143239A1 (en) * 1996-07-18 2006-06-29 Computer Associates International, Inc. Method and apparatus for maintaining data integrity across distributed computer systems
US7085834B2 (en) * 2000-12-22 2006-08-01 Oracle International Corporation Determining a user's groups
US7107543B2 (en) * 2002-01-25 2006-09-12 Tibco Software Inc. Single applet to communicate with multiple HTML elements contained inside of multiple categories on a page
US7127713B2 (en) * 2002-01-11 2006-10-24 Akamai Technologies, Inc. Java application framework for use in a content delivery network (CDN)
US7240101B2 (en) * 2001-04-02 2007-07-03 International Business Machines Corporation Method and apparatus for efficiently reflecting complex systems of objects in XML documents
US7254634B1 (en) * 2002-03-08 2007-08-07 Akamai Technologies, Inc. Managing web tier session state objects in a content delivery network (CDN)

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6173293B1 (en) * 1998-03-13 2001-01-09 Digital Equipment Corporation Scalable distributed file system
US6330561B1 (en) * 1998-06-26 2001-12-11 At&T Corp. Method and apparatus for improving end to end performance of a data network
US7089584B1 (en) * 2000-05-24 2006-08-08 Sun Microsystems, Inc. Security architecture for integration of enterprise information system with J2EE platform
US7571215B2 (en) * 2001-07-16 2009-08-04 Bea Systems, Inc. Data replication protocol
US7568000B2 (en) * 2001-08-21 2009-07-28 Rosemount Analytical Shared-use data processing for process control systems
US20030105837A1 (en) * 2001-11-30 2003-06-05 Yury Kamen Interception for optimal caching of distributed applications
US7403996B2 (en) * 2002-02-21 2008-07-22 Bea Systems, Inc. Systems and methods for migratable services

Patent Citations (54)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5613060A (en) * 1990-05-16 1997-03-18 International Business Machines Corporation Asynchronous resynchronization of a commit procedure
US5261069A (en) * 1990-08-13 1993-11-09 Hewlett-Packard Company Method of maintaining consistency of cached data in a database system
US5581753A (en) * 1994-09-28 1996-12-03 Xerox Corporation Method for providing session consistency guarantees
US5634052A (en) * 1994-10-24 1997-05-27 International Business Machines Corporation System for reducing storage requirements and transmission loads in a backup subsystem in client-server environment by transmitting only delta files from client to server
US5813017A (en) * 1994-10-24 1998-09-22 International Business Machines Corporation System and method for reducing storage requirement in backup subsystems utilizing segmented compression and differencing
US5765171A (en) * 1995-12-29 1998-06-09 Lucent Technologies Inc. Maintaining consistency of database replicas
US5878213A (en) * 1996-02-15 1999-03-02 International Business Machines Corporation Methods, systems and computer program products for the synchronization of time coherent caching system
US6366930B1 (en) * 1996-04-12 2002-04-02 Computer Associates Think, Inc. Intelligent data inventory & asset management systems method and apparatus
US20060143239A1 (en) * 1996-07-18 2006-06-29 Computer Associates International, Inc. Method and apparatus for maintaining data integrity across distributed computer systems
US6086632A (en) * 1996-10-31 2000-07-11 Nec Corporation Register optimizing compiler using commutative operations
US6065046A (en) * 1997-07-29 2000-05-16 Catharon Productions, Inc. Computerized system and associated method of optimally controlled storage and transfer of computer programs on a computer network
US6012059A (en) * 1997-08-21 2000-01-04 Dataxel Corporation Method and apparatus for replicated transaction consistency
US5909689A (en) * 1997-09-18 1999-06-01 Sony Corporation Automatic update of file versions for files shared by several computers which record in respective file directories temporal information for indicating when the files have been created
US6266742B1 (en) * 1997-10-27 2001-07-24 International Business Machines Corporation Algorithm for cache replacement
US6240413B1 (en) * 1997-12-22 2001-05-29 Sun Microsystems, Inc. Fine-grained consistency mechanism for optimistic concurrency control using lock groups
US6542926B2 (en) * 1998-06-10 2003-04-01 Compaq Information Technologies Group, L.P. Software partitioned multi-processor system with flexible resource sharing levels
US6256634B1 (en) * 1998-06-30 2001-07-03 Microsoft Corporation Method and system for purging tombstones for deleted data items in a replicated database
US6304879B1 (en) * 1998-11-25 2001-10-16 Microsoft Corporation Dynamic data cache for object-oriented computing environments
US6453321B1 (en) * 1999-02-11 2002-09-17 Ibm Corporation Structured cache for persistent objects
US6430564B1 (en) * 1999-03-01 2002-08-06 Hewlett-Packard Company Java data manager for embedded device
US6401239B1 (en) * 1999-03-22 2002-06-04 B.I.S. Advanced Software Systems Ltd. System and method for quick downloading of electronic files
US6526521B1 (en) * 1999-06-18 2003-02-25 Emc Corporation Methods and apparatus for providing data storage access
US20010042073A1 (en) * 1999-06-22 2001-11-15 Christian D. Saether Method and system for automatically updating the version of a set of files stored on content servers
US6836889B1 (en) * 1999-08-20 2004-12-28 International Business Machines Corporation Code wrapping to simplify access to and use of enterprise JAVA beans
US6757708B1 (en) * 2000-03-03 2004-06-29 International Business Machines Corporation Caching dynamic content
US20020004850A1 (en) * 2000-03-29 2002-01-10 Krishna Sudarshan System and method of providing a messaging engine for an enterprise javabeans enabled server to achieve container managed asynchronous functionality
US6990526B1 (en) * 2000-05-22 2006-01-24 Pointred Technologies, Inc. Method and apparatus for web caching
US6578160B1 (en) * 2000-05-26 2003-06-10 Emc Corp Hopkinton Fault tolerant, low latency system resource with high level logging of system resource transactions and cross-server mirrored high level logging of system resource transactions
US20040230747A1 (en) * 2000-07-06 2004-11-18 Ims Steven D. Object caching and update queuing technique to improve performance and resource utilization
US6609213B1 (en) * 2000-08-10 2003-08-19 Dell Products, L.P. Cluster-based system and method of recovery from server failures
US6651140B1 (en) * 2000-09-01 2003-11-18 Sun Microsystems, Inc. Caching pattern and method for caching in an object-oriented programming environment
US7000019B2 (en) * 2000-10-17 2006-02-14 Hewlett-Packard/Development Company L.P. Establishment of a deferred network communication session
US6823514B1 (en) * 2000-11-14 2004-11-23 International Business Machines Corporation Method and system for caching across multiple contexts
US20020073188A1 (en) * 2000-12-07 2002-06-13 Rawson Freeman Leigh Method and apparatus for partitioning system management information for a server farm among a plurality of leaseholds
US20020184444A1 (en) * 2000-12-22 2002-12-05 Shandony Michael J. Request based caching of data store data
US7085834B2 (en) * 2000-12-22 2006-08-01 Oracle International Corporation Determining a user's groups
US20020107934A1 (en) * 2001-01-12 2002-08-08 Epicrealm Inc. Method and system for dynamic distributed data caching
US7065616B2 (en) * 2001-02-13 2006-06-20 Network Appliance, Inc. System and method for policy based storage provisioning and management
US20030074580A1 (en) * 2001-03-21 2003-04-17 Knouse Charles W. Access system interface
US7240101B2 (en) * 2001-04-02 2007-07-03 International Business Machines Corporation Method and apparatus for efficiently reflecting complex systems of objects in XML documents
US20030110467A1 (en) * 2001-04-20 2003-06-12 Sree Ayyanar Spinning And Weaving Mills Limited Data storage schema independent programming for data retrieval using semantic bridge
US20020188591A1 (en) * 2001-06-08 2002-12-12 International Business Machines Corporation Disabling and reloading enterprise java beans using database trigger programs
US6918013B2 (en) * 2001-07-16 2005-07-12 Bea Systems, Inc. System and method for flushing bean cache
US20030014480A1 (en) * 2001-07-16 2003-01-16 Sam Pullara Method and apparatus for session replication and failover
US20030050972A1 (en) * 2001-07-17 2003-03-13 Felt Edward P. System and method for transaction processing with transaction property feature
US7028030B2 (en) * 2001-08-30 2006-04-11 Bea Systems, Inc. Cluster caching with concurrency checking
US6826601B2 (en) * 2001-09-06 2004-11-30 Bea Systems, Inc. Exactly one cache framework
US20030065826A1 (en) * 2001-09-06 2003-04-03 Jim Skufca System and method for dynamically caching dynamic multi-sourced persisted EJBs
US20030115366A1 (en) * 2001-12-18 2003-06-19 Robinson Brian R. Asynchronous message delivery system and method
US7127713B2 (en) * 2002-01-11 2006-10-24 Akamai Technologies, Inc. Java application framework for use in a content delivery network (CDN)
US6898587B2 (en) * 2002-01-18 2005-05-24 Bea Systems, Inc. System and method for performing commutative operations in data access systems
US7107543B2 (en) * 2002-01-25 2006-09-12 Tibco Software Inc. Single applet to communicate with multiple HTML elements contained inside of multiple categories on a page
US7254634B1 (en) * 2002-03-08 2007-08-07 Akamai Technologies, Inc. Managing web tier session state objects in a content delivery network (CDN)
US20030200350A1 (en) * 2002-04-19 2003-10-23 Ajay Kumar Class dependency graph-based class loading and reloading

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100161571A1 (en) * 2008-12-18 2010-06-24 Winfried Schwarzmann Ultimate locking mechanism
US8510281B2 (en) * 2008-12-18 2013-08-13 Sap Ag Ultimate locking mechanism

Also Published As

Publication number Publication date
US20030233522A1 (en) 2003-12-18
US7328322B2 (en) 2008-02-05
US7020684B2 (en) 2006-03-28
WO2003063026A1 (en) 2003-07-31
US20060123199A1 (en) 2006-06-08

Similar Documents

Publication Publication Date Title
US7328322B2 (en) System and method for optimistic caching
US8700855B2 (en) System and method for supporting a tiered cache
US6240413B1 (en) Fine-grained consistency mechanism for optimistic concurrency control using lock groups
US6938139B2 (en) Method and system for data element change across multiple instances of data base cache
US7254578B2 (en) Concurrency classes for shared file systems
US5537574A (en) Sysplex shared data coherency method
US7234076B2 (en) Multi-level undo of main-memory and volatile resources
US7831634B2 (en) Initializing a cache region using a generated cache region configuration structure
US7293009B2 (en) System and method for flushing bean cache
US8204931B2 (en) Session management within a multi-tiered enterprise network
US8281014B2 (en) Session lifecycle management within a multi-tiered enterprise network
US9621409B2 (en) System and method for handling storage events in a distributed data grid
US20040158549A1 (en) Method and apparatus for online transaction processing
US20060248285A1 (en) Cache coherence protocol
JPH1063561A (en) System and method for automatically changing database access method to insert database object processing instruction
JP2006012153A (en) Concurrent transactions and page synchronization
US7082432B2 (en) Specifying transaction manager type at various application levels
US20110107338A1 (en) Selecting isolation level for an operation based on manipulated objects
Kotselidis et al. Clustering JVMs with software transactional memory support
Tang et al. Integrating remote invocation and distributed shared state
US20040128328A1 (en) Method and apparatus for relaxed transactional isolation in a client-server caching architecture
Perez-Sorrosal et al. Consistent and scalable cache replication for multi-tier J2EE applications
Niles et al. Exploiting parallelism of distributed nested transactions
Pimentel Self tuning replication in transactional data grids
Ezzio Transactions and Caching

Legal Events

Date Code Title Description
AS Assignment

Owner name: ORACLE INTERNATIONAL CORPORATION, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:BEA SYSTEMS, INC.;REEL/FRAME:025192/0244

Effective date: 20101008

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION