Últimos Cambios |
||
Blog personal: El hilo del laberinto |
Última Actualización: 23 de octubre de 2006 - Lunes
This document was published in the 20061023 release of BerkeleyDB Backend Storage Engine for DURUS.
DURUS is an excellent persistence system for the Python programming language.
This document is not related to the BerkeleyDB storage engine for Durus, but tries to clarify Durus operation and inner working.
This document describes operation of Durus 3.5 release.
If you find an error in the documentation, or would like it to be expanded to include new topics, send a mail to jcea@jcea.es.
For ease navigation in this document, each section begins with three "###". You can use such a sequence to go around this text. Each section, also, documents the date of last modification.
### Concurrency using the Durus Storage Server (20060421)
The Durus Storage Server allows a single storage to be shared between several (remote) clients. So you can access the storage remotely, and writes by a client will be visible to others.
Durus Storage Server will listen for requests from all clients connected, but when any request arrives, the server will be busy ONLY with that client alone. Other requests will be queued until finished. If that client is very slow or the disk access is slow, the server will sit idle, even if other clients are demanding attention.
Hope a future Durus release can process multiple read requests in parallel. Each client would wait less, and the disk would be better utilized (better sort multiple seeks to serve several request that a long seek to serve only one request).
Remember, nevertheless, that Durus clients have a local cache to avoid hitting the storage server. Sizing that cache, and learning how to use it in an effective way, are important issues in any demanding Durus deployment.
### ACID using the Durus Storage Server (20061023)
ACID = Atomicity, Consistency, Isolation, Durability
DSS = Durus Storage Server
Since DSS only processes a request from a single client at a time, commits are atomic. No other client will be served until the commit completes.
The Durability is garanteed by the Storage Backend used by Durus. Some backends (for example, my BerkeleyDB Storage backend) can be configured to not garantee Durability in exchange of (vastly) improved performance. Some applications can take advantage of that. Some other requires durability.
Transactions under DSS are Isolated. If you don't do any dirty trick, DSS garantee a "degree 3 isolation". That is, you only see committed data, and reads are repeatable.
Consistence is provided also by the Storage Backend used by Durus. It implies that no transaction can leave the Storage in an inconsistent state physically. If the application logic has integrity constraits, it must be enforced by the application.
### Durus Storage Server conflicts (20061023)
Durus clients implement a local cache to improve performance, avoiding DSS accesses. Objects fetched or written are kept in a cache. The cache size is configurable, and evictions are transparent.
The eviction routine can be directly called or, better, automatically done when you do a transaction commit or abort. Cache consistency is also checked when you fetch objects not present in you cache.
Cache consistency is checked when you do a commit or abort. Each connection has its own cache, even in the same process.
If you do an abort, locally modified objects are purged. If the cache has objects that other client modified, they are also purged. So, after an abort, your cache only keeps unmodified objects, both locally and remotely.
If you do a commit, it will fail if you touched any object also remotely modified by another client, during your current transaction. If your commit conflicts, the eviction procedure will be like the abort case.
If your commit successes, your cache will purge remotelly modified objects not used in current transaction.
If your code touches an object not in cache, and that object was modified remotely since your current transaction started, you will get a conflict in that moment, before the commit/abort.
Some discussion about this issue:
http://mail.mems-exchange.org/durusmail/durus-users/508/
http://mail.mems-exchange.org/durusmail/durus-users/514/
Another important issue is that DSS keeps a changeset per connected client, with the OIDs of the objects changed (and commited) by the other Durus clients. That changeset is sent (and then cleared) to its client when it does a commit or an abort, in order to synchronize its cache. This system has two consequences:
If the "idle" client has a very low "duty" cycle, could be better to simply break the DSS connection.
The changeset size depends of the number of objects changed, and the change rate. But if you have a lot of writes to a small object set, changeset size will be small. It tracks what objects are changed, not how many changes where done to an object.
http://mail.mems-exchange.org/durusmail/durus-users/379/
Also, keep your transactions as sort as you can, to reduce conflict risk.
The usual approach to conflicts is to abort the transaction, repeat the computation with up-to-date data, and try again. All the times you need. Technically, this transactional approach is called "optimistic concurrency", and it is a standard algorithm in database world, mainly used when conflicts are rare.
### Persistent object references and transactions (20060520)
Keeping around persistent object references between transactions is calling for trouble. You SHOULDN'T do that.
Your references can get stale without notice, specially using Storage backends like my BerkeleyDB Storage backend, that deletes garbage objects promptly. Example:
http://mail.mems-exchange.org/durusmail/durus-users/397/
Other problems with keeping references around are:
You should always access objects from "root" object, without keeping intermediate references, unless you know what are you doing and the inner lifecycle of those objects. You can only safely keep references while in a transaction, but discard them at commit/abort time.
You can keep weak references to loaded objects, nevertheless, across transaction boundaries. More info later.
About this section, David K. Hess <dhess@verscend.com> tells me in a private email:
If you are using FileStorage rather than ClientStorage, then (at least it seems in my testing and usage) persistent references to objects are not only safe but handy. When combined with a regular call to commit(), it becomes a very clean, easy and efficient way to persist a singe application's internal data structures to disk. This results in having a snap shot that is very easy to reload application state from in case of a crash/shutdown.
Yes, if a program is the only "mutator" to its data structures, then you don't need to travel all the path from the root, if you are sure that a given object has no vanished under your foot.
### Durus clients, threading and multiple connections (20061023)
BEWARE: Stock Durus 3.5 has several issues with threads. Please, apply patch in http://mail.mems-exchange.org/durusmail/durus-users/698/ This patch is supposed to be included in a future Durus release.
Durus is not threadsafe, but you can use it in a threaded python program if you take care:
This is critical, since DSS accesses from multiple threads could be intermixed and a crash would be the better outcome (you would crash the DSS, also, and corrupt your storage database).
You can't coordinate transactions between different DSS connections.
Since sharing objects is forbidden, you can only exchange state between connections, even in the same program, going tru the DSS, using normal transactions.
Don't do that unless you know what are you doing.
### Data compression in the Storage (20061023)
By default Durus stores object data compressed on disk. The algorithm used is zlib (http://www.zlib.net/)
In some situations compression could be inconvenient. For example, the data is big and already compressed (let say, graphics). Or, perhaps, a better algorithm could be used with that data.
You can disable compression simply setting "WRITE_COMPRESSED_STATE_PICKLES" to False in "durus.serialize". This way Durus will save new and modified objects uncompressed. Durus will load correctly both compressed and uncompressed objects, nevertheless, so you don't need to update all your database.
If you need to personalize your compression, you can follow advice by David Binger: (http://mail.mems-exchange.org/durusmail/durus-users/492/)
Here's what I would do. Set WRITE_COMPRESSED_STATE_PICKLES to False. Add __getstate__() and __setstate__() methods to your persistent classes that provide the customized compression behavior. If you want compressed pickles for a certain class, make the __getstate__() return a compressed pickle of self.__dict__ instead of the dict itself. The __setstate__() must have the corresponding inverse behavior.
A curiosity note: all zlib streams starts with the char "x", so if your "__getstate__()" returns a string starting with "x" when loading, Durus will try to unzip it. It will fails, of course, and then your "__setstate__()" will be called. So, if you are worried about efficiency, be sure your "__getstate__()" strings never starts with a "x" char :-).
### Weak references (20061023)
Your program can keep internal (non persistent) weak references to persistent loaded instantes. Those references will be cleared automatically if necessary: cache shrink, conflicts, etc. You can keep such references across transactions boundaries.
Persistent objects can't have weak references to other persistent objects to ease garbage collection in the Storage. All interobject references in the Storage will be strong.
Nevertheless you can simulate sort-of weak references by hand using the internal OID of referenced persistent objects, since you can use the connection's "get()" method to load a persistent object given its OID. This manually managed reference doesn't preclude garbage collection of referenced objects if their reference count goes to zero. Just like standard weak references.
Some details:
### Implicit object loading/dump (20061023)
Transparent object loading/dumping is the key to a successful persistence system. The details are simple to understand when you "get it". You can read more about this in:
http://mail.mems-exchange.org/durusmail/durus-users/533/
Some random details:
So if you load a persistent object with references to 1000 other persistent objects, only the state of the parent object will be loaded, but 1000 ghost objects will be created.
Remember also that "__init__" will be called only when the object is created first time, not each time it is loaded in RAM.
### "gen_oid_record()" (20060424)
Durus storages backends usually define a "gen_oid_record()" method. That method iterates over all the objects in the Storage, in no particular order. Current backend implementations have the following caveats: (http://mail.mems-exchange.org/durusmail/durus-users/500/)
This method is usually used to convert a storage to other format, or to update classes of already stored objects. You can use it, also, for backup purposes.
The usual approach is to iterate over the source storage, loading objects, and storing them as-is in the destination storage. When the migration is done, you do a "giant" commit. This approach is doable when your database is small enough to be loaded in RAM+SWAP but if your machine is 32 bits, you are ultimate limited by the addressable space you have, typically in the 2^30 bytes order.
You can't do multiple smaller "commits" because some storages (for example, my BerkeleyDB storage backend implementation) would do a background garbage collection and delete copied but not yet referenced objects.
Releases of my BerkeleyDB storage backend from 20060509 includes a "migrate()" method to migrate HUGE datasets with no RAM requirements, in a clean and efficient way.
Remember also that "gen_oid_record()" in the ClientStorage (the standard DSS implementation) is very inefficient. Time to transfer all the objects will be O(MAX_OID) and not to O(N). That is, time will be proportional to the number of OIDs ever generated, not to the number of really existant objects.
### ComputedAttribute (20061023)
ComputedAttribute's are especial persistent classes without state, used to keep (in RAM) cached values of "costly" functions. That cached values are discarded if the instance is purged from memory (for instance, cache shrink) or if any other DSS client sent an "invalidation".
The access to the cached value is done via a "get()" method. If the cached value is still current, we will get it. If the cached value wasn't computed before, or was invalidated, a new value will be computed and cached.
The function used to compute the cached value, if necesary, is passed as a parameter to the "get()" method. That function MUST NOT take any parameter. This seems to be a huge issue, but you can use a lambda or closures to pass "hidden" parameters. In python 2.5 and up, you could also use partial function application.
Some comments:
### Non persistent attributes (20060424)
Durus has no support for non-persistent attributes. That is, all attributes are stored on disk, ever.
See: http://mail.mems-exchange.org/durusmail/durus-users/411/
I guess you can implement them in your own persistent classes touching "__getstate__".
Keep in mind comment from David Binger:
In my opinion, convenience attributes on Persistent instances like this invite trouble.
### Newly created persistent objects and conflicts/aborts (20061023)
When you create a new persistent instance, it is not associated to a particular connection, so transaction related actions (commits, aborts, conflicts) don't affect the new object.
When you link your new persistent object to an already persistent object you have, you are linking the new object to the connection vinculated to that old object. Now you have three cases:
If the object is not touched, you can reuse it as-is in a new transaction try. You don't need to "recreate it", unless you had a conflict and the data in the new object was based in stale objects. Of course, in that case you must recalculate the data.
If the object is not vinculated to a connection, you can transfer it to another one or to another thread. That is, "free" new objects can be shared between threads. But only while the new object is not vinculated to a particular connection via a link from another persistent object.
Since Durus 3.5, Durus tries to reuse OIDs when a conflict arises, to save OIDs (not an issue actually, since you have 2^64 available) and to abreviate network latency issues. From this version, also, OIDs are requested in a block, for far improved commit time when commiting a lot of newly created persistent objects.
### BTree methods (20061003)
Durus provides several persistent classes to use in your programs. The most interesting and "different" is BTree.
BTree provides a (persistent) dictionary-like class. The main advantage is that a BTree is not fully loaded in RAM. Only the elements needed are fetched. So you can have an arbitrary huge BTree, without eating your RAM.
As said, BTree is used like a normal python dictionary, but there are some additional useful features:
Más información sobre los OpenBadges
Donación BitCoin: 19niBN42ac2pqDQFx6GJZxry2JQSFvwAfS