NAME OOPS - Object Oriented Persistent Store SYNOPSIS use OOPS; transaction(sub { $oops = new OOPS dbi_dsn => $DBI_DSN, username => $username, password => $password, table_prefix => "MY", readonly => 0; $oops->commit; $oops->{my_1st_bit_of_data} = 'a scalar'; $oops->{my_2nd_bit_of_data} = { A => 'hash' }; $oops->{my_3rd_bit_of_data} = [ qw(An Array) ]; $oops->{my_4rd_bit_of_data} = \'a reference to a scalar, ref, hash, or array' ]; my $old_value = $oops->{multiple}{level}{dereference}; my $object = $oops->load_object($integer_object_id); my $dbh = $oops->dbh(); $oops->workaround27555($reference); my $was_virtual = $oops->virtual_object(\%{$oops->{some}{hash}, [ $new_value ]); }); my $ref = getref(%hash, 'key') DESCRIPTION The goal of OOPS is to make perl objects easily and transparently persistent. OOPS handles deeply nested and cross-linked objects -- even object hierarchies that are too large to fit in memory and (with a hint) individual hash tables that are too large for memory. Objects will be demand-loaded into memory as they are accessed. All changes to your object hierarchy will be saved with a single commit(). Full transactional consistency is the only operational mode. Either all of your changes are saved or none of them are. While your program runs, you will see a consistent view of the data: no other running transactions will change the data you see. If another transaction changes data that you are using then at least one of the transactions must abort. OOPS will die() to abort the transaction. OOPS maps all perl objects to the same RDBMS schema. No advance schema definition is required on the part of the user of OOPS. The name of the package (OOPS) comes from the realization that perl's data model is much more complicated than I initially understood. Internally, the RDBMS schema uses four tables: a table of objects, a table of attributes (keys and values), a table of large attributes that are too big to fit in normal (VARCHAR) columns, and a table of counters. At this time, OOPS is expecting a web-like work flow: create OOPS instance access some objects modify some objects commit exit If you need more than one transaction in a program, create more than one OOPS instance. To make your data persistent, make a reference to your data from the OOPS object. To later retrieve your data, simply access it through the OOPS object. EXAMPLE PROGRAM use OOPS; transaction(sub { my $oops = new OOPS dbi_dsn => 'DBI:mysql:database=MY-DATABASE-NAME;host=localhost', username => 'MY-USERNAME', password => 'MY-PASSWORD', table_prefix => "MY-TABLE-PREFIX"; my $p = $oops->{pages}{"/some/path"}; $p->{next} = $oops->{pages}{"/some/other/path"}; $p->{jpgs} = [ read_file("x.jpg"), read_file("y.jpg") ]; $oops->commit; }); exit; SUPPORTED DATA TYPES Perl blessed and unblessed HASHes, REFs, SCALARs, and ARRAYs are supported. Currently, HASH keys may not be longer than 255 characters. Class names may not be more than 128 characters long. References to hash keys and array elements are supported. At the current time, large ARRAYs are not efficient. Use HASHes instead if this matters to you. References to array elements and hash values are not efficient. Large HASHes are supported by only loading keys as tehy are accessed. HASHes, array elements, and REFs are implemented with tie(). ARRAYs are not currently tie()d because of bugs in perl. Multiple references to the same scalar are supported. References to array elements and hash values are supported. Persistent data is reference counted and cycles must be manually broken to assure de-allocation. A persistent data garbage collector OOPS::GC is supplied if you cannot guarantee to break your own cycles. SUPPORTED PLATFORMS Support for the following RDBMSs are is included: PostgreSQL OOPS 0.2005 has been tested with PostgreSQL version 9.1.3 on Linux. OOPS 0.2002 has been tested with PostgreSQL version 8.1.8 on Linux. OOPS 0.1004 was tested with PostgreSQL version 8.1.4 on Linux. OOPS 0.1003 was tested with 7.4.2 and 7.3.5 on Linux and 7.3.5 on pre-release DragonflyBSD. Somewhere between 7.4.2 and 8.1.4 PostgreSQL got more strict about TEXT versus BYTEA types. OOPS 1.003 was using TEXT where it should have been using BYTEA. Be sure to upgrade OOPS to 0.1004 or later *before* upgrading PostgreSQL to 8.x or later. Biggest issue: PostgreSQL runs the test suite more slowly than mysql or SQLite. On rare occaision PostgresSQL will fail with an "ERROR: duplicate key violates unique constraint" error. If this happens inside a call to "transaction()" it will automatically restart. You only need to worry about this if you are not using "transaction()". mysql OOPS 0.2005 has been tested with mysql 5.1.46 on Linux. OOPS 0.2003 has been tested with mysql 5.0.45 on Linux. OOPS 0.2002 has been tested with mysql 5.0.22 on Linux. OOPS 0.1004 has been tested with mysql 4.1.15 on Linux. OOPS 0.1003 was tested with mysql 4.0.16 and 4.0.18 using InnoDB tables. The OOPS schema for mysql does not support large (>256 byte) hash keys. Mysql locking doesn't work correctly. To get correct behavior, access to the database must be exclusive . This problem is fixed in 5.0.45. SQLite2 DBD::SQLite versions 0.x use SQLite 2.x. DBD::SQLite2 is the continuation of DBD::SQLite version 0.x codebase and uses SQLite 2.x. DBD::SQLite2 is not 8-bit clean with respect to "\0". OOPS uses a DBD::SQLite2 feature to translate binary nulls. A side-effect is that backslash will be doubled "\" -> "\\" in DBD::SQLite2 (the extras will be removed transparently by DBD::SQLite2. OOPS does not use the "counters" table with DBD::SQLite2. SQLite DBD::SQLite versions 1.x use SQLite 3.x. DBD::SQLite versions 1.x leak file descriptors and will eventually die if you create use and destroy too many OOPS instances. DBD::SQLite versions 1.x cannot pass the regression suite and may not be reliable in production use. Use DBD::SQLite2 instead. OOPS does not use the "counters" table with DBD::SQLite. SQLite is not 8-bit clean with respect to "\0". OOPS uses a DBD::SQLite feature to translate binary nulls. A side-effect is that backslash will be doubled "\" -> "\\" in DBD::SQLite (the extras will be removed transparently by DBD::SQLite. As of 0.2002, OOPS doesn't make use of the newer (version 1.x) DBD::SQLite features for handling nulls. To work around the problem with the leaked file descriptors, on systems that support "/proc", file descriptors in excess of $OOPS::sqlite_v3::enable_fd_leak_workaround (500 by default) that match the sqlite database file will be closed when new OOPS instances are created. This sometimes helps. Set $OOPS::sqlite_v3::enable_fd_leak_workaround to 0 to disable. Related bug reports: , . Perl versions 5.8.2 through 5.12.1 are supported. Prior to 5.8.2, it wasn't possible to untie scalars from within the a tied scalar access method. An ugly workaround is possible if there is enough interest. OOPS 0.2002 and 0.1004 have been tested on Linux 2.6.15 (Ubuntu 6.06LTS - Dapper Drake). OOPS 0.1003 was tested on Linux 2.4.23 (Debian unstable as of April '04); on FreeBSD 4.9; and on DragonflyBSD prerelease. As far as performance goes, mysql and SQLite are both about twice as fast as PostgreSQL for applications that only have one transaction at at time. SQLite and mysql are particularly slow when there are multiple transactions as their lock granularity is the entire database. Each RDBMS uses a somewhat different schema. FUNCTIONS "transaction($funcref, @args)" "transaction()" is a wrapper for a complete transaction. Transactions that fail due to deadlock with other processes will be re-run automatically. The first parameter is a reference to a function. Any additional parameters will be passed as parameters to that function. The return value of "transaction()" is the return value of "&$funcref()". It is not necessary to use the "transaction()" method. Beware that nearly any operation on persistent date (even read operations) can cause deadlock. Any use of persistent data can trigger a deadlock. The "transaction()" function catches this and retries automatically up to $OOPS::transaction_maxtries times (15 times unless you change it). If you don't use "transaction()" you might want to catch the exceptions that "transaction()" catches. To do this, you can regex match $@ against $OOPS::transfailrx. Basically, transaction is a slightly fancier version of the following: sub transaction { my ($sub, @args) = @_; for (0..15) { eval { &$sub(@args); } last unless $@; next if $@ =~ /$OOPS::transfailrx/; die $@; } } The important thing to notice is that your code will be called over and over until the transaction succeeds. This means you should write your code so that it doesn't have any external side effects until after it calls commit(). The following package variables are available during transactions: $OOPS::transaction_tries The number of attempts at this transaction. Starts with 1. $OOPS::transaction_maxtries The maximum number of attempts that will be made. $OOPS::transaction_failure_sleep The minimum amount of time to "sleep()" if the transaction fails. $OOPS::transaction_failure_maxsleep The maximum amount of time to "sleep()" if the transaction fails. @OOPS::transaction_rollback A list of functions to call in the event of a transaction rollback (code refs). "getref(%hash, $key)" References to tied hash keys are buggy in all perls through 5.8.7 (and beyond?). Use "getref(%hash, $key)" to create your reference to a tied hash key. See: and . $ref = getref(%hash, $key); Alternatively, use "$oops->workaround27555($ref)". Getref() and workaround27555() work around all the perl bugs with tied hash key references. Failure to use them may result in unexpected and inconsistent results. "walk_hash(%hash, $stride, [ $previous_last_key ])" If you have a large virtual hash, you may want to be able to iterate over the keys in the hash as a bunch of transactions rather than as one transaction. Walk_hash() can be called on any HASH. If it isn't an OOPS virtual hash, it will return keys in batches just like if it had been called on an OOPS virtual hash. Here is an example of how to use it: my $stride = 20; my $done = 0; my $lastkey; while (not $done) { transaction(sub { my $oops = OOPS->new(...); my @things = walk_hash(%{$oops->{my_table}}, $stride, $lastkey); for my $thing (@things) { # do stuff } $lastkey = $things[$#things]; $done = 1 unless @things == $stride; }); } Here is another example of how to use it (with OOPS::TxHash): use OOPS::TxHash; my $stride = 20; my %real_todo = ( stuff => undef ); while (%real_todo) { transaction(sub { my $th = tie my %todo, 'OOPS::TxHash', \%real_todo or die; my $oops = OOPS->new(...); if (exists $todo{stuff}) { my @keys = walk_hash(%{$oops->{stuff}}, $stride, $todo{stuff}); for my $k (@keys) { # work on $oops->{stuff}{$k} } $todo{stuff} = $keys[$#keys]; delete $todo{stuff} unless @keys == $stride; } $oops->commit; $th->commit; }); } Here is yet another example (with a wrapper function): walk_eval(20, sub { OOPS->new(...) }, sub { my $oops = shift; $oops->{my_table} }, sub { my ($oops, $thing) = @_; # do stuff }, allatonce => 0 ); sub walk_eval { my ($stride, $get_oops, $get_hash, $agent, %opts) = @_; my $done = 0; my $last = undef; $stride ||= 100; while (! $done) { transaction(sub { my $oops = &$get_oops(); my $hash = &$get_hash($oops); my @items = walk_hash(%$hash, $stride, $last); if ($opts{allatonce}) { &$agent($oops, @items); } else { for my $item (@items) { &$agent($oops, $item); } } $oops->commit(); $last = $items[$#items]; $done = 1 unless @items == $stride; }); } } PUBLIC CLASS METHODS "OOPS->new(%parameters)" Creates a OOPS object instance. More than one object instance is allowed at the same time. Instances are unaware of each other. Making a reference from a persistent object in one instance to a persistent object in another instance will cause recursive copying from one instance to the other. (Untested). The %parameters are documented in the next section. "OOPS->initial_setup(%parameters)" Drops and recreates the database tables. Don't use it too often :-) The regression suite drops and re-creates the tables many times. The %parameters are documented in the next section. PARAMETERS FOR NEW The "new()" and "initial_setup()" methods take a hash specification for their behavior. Here are the parameters allowed: "dbi_dsn" / $ENV{OOPS_DSN} / $ENV{OOPS_DRIVER} / $ENV{DBI_DSN} / "$ENV{DBI_DRIVER" Many ways to specify the DSN for DBI: as an argument; as an OOPS environment variable; as a DBI environment variable. Use at least one :-) See the DBI documentation for more details. "user" / $ENV{OOPS_USER} / $ENV{DBI_USER} Three ways to specify the user for DBI: as an argument; as an OOPS environment variable; as a DBI environment variable. Not required for all databases. "password" / $ENV{OOPS_PASS} / $ENV{DBI_PASS} Three ways to specify the password for DBI: as an argument; as an OOPS environment variable; as a DBI environment variable. Not required for all databases. "readonly" If true, commit() will be disabled and the transaction isolation level will be set to READ COMMITTED instead of SERIALIZABLE for the databases where this matters (mysql, PostgreSQL). This should be used if you plan to scan across large amounts of data because otherwise everything you read will end up locked. This feature is still experimental (not covered by the test suite). "less_caching" If true, iterating over the keys of a virtual hash (see "virtual_object") will not cache the key values in memory. If your virtual hash is too big to fit in memory, this is a very good idea. Alternatively, you can use the "clear_cache" method (below). This feature is still experimental (not covered by the test suite). "table_prefix" / $ENV{OOPS_PREFIX} OOPS allows a prefix to be supplied for it's internal table names. If you set a prefix of "FOO_" then it will use a "FOO_object" table instead of an "object" table. This can be set as an argument to "new()" or it can be set with the environment variable $OOPS_PREFIX. This allows multiple separate object spaces to exist within the same backend SQL database. It's intended use is to support testing vs. production environments but it could also be used to segregate object spaces that don't link to each other. The current tables and indexes are: object Each row is an object. attribute Each row is an attribute value in an object. big Each row is a large attribute value. counters Counters for things like object ids. temp A temporary table. group_index Object grouping index. value_index Lookup by attribute value. "auto_upgrade" / $ENV{OOPS_UPGRADE} The relational schema for OOPS is not same for all versions of OOPS. If this is set, then an older schema will be upgraded to the current schema. If this is not set, then OOPS will use the older schema unchanged. In most cases, this means that OOPS will use a historical version of itself rather than the current version. See the SCHEMA VERSIONS section of this document. "auto_initialize" / $ENV{OOPS_INIT} The DBMS must be initialized before it can be used: tables created, a few rows inserted, etc. This can be handled by calling "initial_setup()" or it can be handled automatically by "new()" if "auto_initialize" or $OOPS_AUTO_INIT is set. "default_synchronous" / $ENV{OOPS_SYNC} With SQLite, an additional parameter to "OOPS->new()" is recognized: "default_synchronous". Possible values are: FULL Sync() all transactions to disk before returning. NORMAL The default: sync() at critical moments only - protects against program failure, but not all power or OS failures. OFF Don't sync() at all and go really fast. PUBLIC OBJECT METHODS "->commit()" Writes any changed objects back to the database and commits the transaction. Currently only one commit() call is allowed. Do not access your persistent data after commit() -- it may work but this is not covered well in the regression suite. "->virtual_object(\%hash [,$new_value])" Queries [and sets] the load-virtual flag on a persistent hash. Hashes that load virtual will do separate queries for each key rather than load the entire hash. This is a good thing if your has has lots of keys. This flag takes effect the next time the hash is loaded. The value is a perl boolean. This may be handled automatically in the future. "->workaround27555($reference)" References to tied hash keys are buggy in all perls through 5.12.1 (and beyond?). Use workaround27555($reference) to register your new tied hash key references so that they can be transformed into references that actually work correctly. $ref = \%hash{$key}; $oops->workaround27555($ref); "workaround27555()" is harmless if called on other sorts of references so it is safe to use indiscriminately. See . Alternatively, use "getref(%hash, $key)". "->dbh()" This returns the main DBI database handle used by OOPS. This function is provided for those who want to hand-write queries. Please note: no changes are written to the DBMS by OOPS until "$oops-"commit()>. "->load_object($id)" This will load a persistent object by number. It returns the object or undef if the object doesn't exists. This function is provided for those who want to hand-write queries. "->clear_cache()" This will clear out some of the cache that oops maintains. It is useful if you're iterating over the keys of a large virtual object and don't want to overflow memory. This feature is still experimental (not covered by the test suite). "->lock($ref)" Lock an object or a scalar. While this is never needed for transactional consistentcy within the OOPS data, it can be useful for transactions that have external side-effects like sending an email. In theory, if you pre-lock everything you might change, your transaction will not need to be rolled back due to deadlock. When you lock an object, it is the object record that is locked, not the contents. Locking "\%hash" won't stop another process from modifying "$hash->{key}". Lock "\$hash->{key}" if you want that. LIMITATIONS, ERRATA, DEVELOPMENT STATUS, BUGS OOPS has been thoroughly tested. The regression suite is very well developed: there is twice as much code in the test suite as there is in the module itself. The suite does over 1.5million tests. I have so much confidence in my test suite, I'm offering a bounty on bugs! Limitations OOPS caches in memory all data that is pulled in during a transaction. For "virtual_object" hash'es only, there are ways to prevent absolutely everything from being cached (the "less_caching" contructor agrument and the "clear_cache" method) but for other objects you need to have enough memory to hold everything you touch. You can only call ->commit once per OOPS instance. The cache is emptied when the OOPS instance is destroyed. This means that OOPS will be loading up its cache over and over. Known bugs in OOPS memory leaks OOPS currently has memory leaks. This may or may not matter to your application. The rate of leakage varies depending on which RDBMS is used. SQLite seems to have the most significant problems. Most of the leaks are not in OOPS itself but in the modules it uses and thus are not easily addressed by changing OOPS. delayed DESTROY Additional references to the in-memory copies of persistent data are kept by OOPS. These extra references will prevent DESTROY methods from being called as soon as they otherwise would be. They'll usually be delayed until the OOPS object is itself DESTROYed. other magic Other perl magic attributes are not currently stored persistently. Many probably could be supported, but many could not. For example, taint does not work on tied hashes: . unreferenced blessed scalars When you bless a reference to a scalar value, the blessing is stored with the scalar, not the reference. The blessing remains even if there is no reference to the scalar. The following code prints "true". my $x = 'foobar'; my $y = \$x; bless $y, 'baz'; $y = 7; $y = \$x; my $z = ref($y); print "true\n" if $z eq 'baz'; At the current time, OOPS does not store such blessings. OOPS does not remember blessings when there isn't a reference. re-blessing the OOPS object Reblessing the OOPS object is not supported. DBD::Pg and ASCII NULL DBD::Pg does not easily support ASCII NULL. OOPS has only partial support for ASCII NULL with PostgreSQL. Everthing works, but don't have ASCII NULL in your class names. Circular references sometimes break OOPS mostly allows you to make circular self-references: my $x; $x = \$x; In some (rare) cases circular self-references sometimes break. Since I can't think on any good reason to make a circular self-reference, I'm not going to fix this bug. Long hash keys The database schema for OOPS does not support huge hash keys. Bugs in perl that effect OOPS References to hash keys Persistent hashes are implemented with tie. There are bugs with perl's implementation of references to tied hash keys. These bugs will be triggered in several situations: creating a reference to tied hash key that doesn't exist yet; deleting a key that has a reference tied to it; assigning through a reference to a key that has multiple references. All of the above can either be avoided or you can workaround them by either calling "workaround27555"($YOUR_REFERENCE) whenever you create a tied hash key reference or by using "getref(%hash, $key)" to create your reference. The perl bugs are documented in: and . "local" and tie "local(%some_tied_hash)" doesn't work right. Thus "local(%some_persistent_hash)" won't work right either: . "scalar(%hash)" Tied "scalar(%hash)" support was added in perl 5.8.3 and does not exist in 5.8.2. Tied arrays don't work right There are a couple of bugs with tied arrays that prevent OOPS from using them: and . OOPS fully loads arrays into memory to work-around this problem. This isn't a big deal unless you've got big arrays. This was fixed in perl 5.10, but OOPS does not yet take advantage of tied arrays. SQLite and perl's malloc(). If SQLite is used with a perl that has been compiled to use perl's "malloc()", it will report LOTS of "Bad free() ignored (PERL_CORE)" errors. It is not currently known if these errors are harmful beyond generating lots of output to STDERR. The default perl configuration on FreeBSD uses perl's "malloc()". SCHEMA VERSIONS As OOPS is developed, the database schema that OOPS uses changes. OOPS notices if the version of the schema in the DBMS is different than the version the code currently supports. When this happens there are two possibilities: either OOPS will upgrade the schema to the current version or it will use an older version of OOPS to access the data. See the "auto_upgrade" notes in the PARAMETERS FOR NEW section. This behavior allows the OOPS module to be upgraded without disrupting installed applications. DATABASE ADMINISTRATION Like Perl, OOPS uses reference counting. If you have circular data structures and you delete all you references to the circular data structure, you'll leak persistent memory. Use OOPS::GC to mop up the leak. Previous versions of OOPS have had bugs that caused the reference counts on the persistent data to become inaccurate. Use OOPS::Fsck to fix such problems. Do this before once, before using OOPS::GC. FUTURE DIRECTIONS OOPS isn't done. There are a bunch of things that I am considering adding to it. If any of these things is important to you, speak up so that I know there is interest... fix the bugs There are bugs listed in the DEVELOPMENT STATUS section that could be fixed. First up is fixing the memory leaks that are in OOPS itself. code cleanup and general performance enhancements The initial releases of OOPS concentrated on correct behavior and other aspects of the module were somewhat ignored. The code could be cleaned up a bunch. perl-syntax SQL query translator SELECT Employee WHERE $Employee->{salary} > 5000 It's possible to translate perl-syntax queries into real SQL that can be used to query the object store. better grouping Objects are loaded in groups rather than individually. There is much room for improvement in choosing how groups are formed. This is largely undeveloped as yet. caching Many possibilities. A cache-invalidation daemon to note when objects have changed. Re-verification of touched data from the database. Ability to call commit() more than once. weak references Support for persistent weak references is possible. external references to objects Currently objects are reference counted internally. You must have a reference to something from an already existing object for it to continue to exist. contracts OOPS has to do a lot of scanning of objects to see if they've changed. Explicit notification of changes would improve performance. OOPS could call functions before saving and after loading to transform objects for a better or cleaner on-disk representation. support for 'base' & accessor methods This isn't something that I care about but maybe someone else does? schema enforcement Allow explicit schemas to be defined. Do not save objects that don't conform. Eg: Hash::Util::lock_keys(). RDBMS -> object mapping Map existing RDBMS schemas into objects. data viewer Viewing large datasets of deep and cross-linked data is difficult. Perhaps a CGI-based or Tk-based data navigator would help. support for tied data structures It is possible to support storing tied data. The tied object is what would need to be persistent. This would only work on some kinds of ties. support for other base types. Right now, just HASH, SCALAR, REF, and ARRAY are supported. Regular expressions, file handles, I don't know it's possible to support code references. on-line data migration By doing double updates, OOPS could support live migration from one DBMS to another. WRITING SQL QUERIES BY HAND If you want to query your data, then until a translator is written, your only choice for making queries is to write them by hand. Using your data does not require a query: anything you've got a reference to will be loaded as you access it. Queries are for performing searches that don't have a perl-object index. Each perl HASH, REF/SCALAR, or ARRAY has a row in the "object" table and multiple rows in the "attribute" and "big" tables. Here are the columns you'll care about: object There is one row per perl object. id The object id. class The blessed class name (limited to 255 characters). otype The type of object: 'H' A HASH. 'A' An ARRAY. 'S' A SCALAR or REF. attribute This is a table of key/value pairs. The keys correspond to perl hash keys and perl array indexes. The values correspond to perl hash values and array element values. id The object id. pkey The hash key or array index. pval he hash value or array value. Limited to 255 characters. ptype Flags the type of the value. Possible values are: '0' A normal value. Numeric or string. 'B' An big value. "pval" will be a copy of the start of the value for the first N characters. The end of "pval" will be a MD5 checksum of the full big value. 'R' A reference to another object. big This is a table of values that were too large for the normal columns. Even with databases that support wide columns, a separate big table is used so that you don't load large scalars unless you actually need the value. id The object id. pkey The hash key or array index. pval The hash value or array value. Limited to whatever the underlying database will support as it's largest blob. fragno Blob fragment number. This column only exists with SQLite. SQLite has a smallish maximum row size and so big values must be split into multiple rows. REFs are are special. There are several types of REFs: references to scalar values; references to objects; secondary references to scalar values; references to scalar values that are part of another object (references to hash keys and references to array elements). The representation of references is designed so that you don't need to care what sort of REF it is when you're doing a query. The basic REF is a ref to a value inside another object. An example: OBJECT TABLE id class otype 1 OOPS::NamedObj H 383 SCALAR R 384 SCALAR R 385 SCALAR R 386 REF R 400 HASH H 500 ARRAY A ATTRIBUTE TABLE id pkey pval ptype 1 A500 500 R 1 H400 400 R 1 R383 383 R 1 R384 384 R 1 R385 385 R 1 R386 386 R 383 400 'a-key' 0 384 384 'nopkey' 0 384 'nopkey' 'a-value' 0 385 384 'nopkey' 0 386 386 'nopkey' 0 386 'nopkey' 500 R 400 'a-key' 'a-value 0 400 'another-key' 'another-value' 0 400 'A500' 500 R 500 0 'a-value' 0 500 1 'another-value' 0 HASH 1 is %$oops. REF 383 is a reference to the key 'a-key' in object #400 (a HASH). REF 384 is a ref to scalar. It uses two rows to make writing queries easier. REF 385 is a duplicate reference to a scalar value. It duplicates REF 384. In behavior, these two REFs should be identical even though they are represented differently in the database. REF 386 is a ref to an object: #500 (an ARRAY). HASH 400 is a normal hash. ARRAY 500 is a normal hash. This example data is what you would end up with after running code like: my $oops = new OOPS dbi_dsn => 'DBI:mysql:database=MY-DATABASE-NAME;host=localhost', username => 'MY-USERNAME', password => 'MY-PASSWORD'; $oops->{A500} = [ 'a-value', 'another-value' ]; $oops->{H400} = { 'a-key' => 'a-value', 'another-key' => 'another-value', 'A500' => $oops->{A500}, }; $oops->{R383} = \$oops->{H400}{'a-key'}; $oops->workaround27555($oops->{R383}); $oops->{R384} = \'a-value'; $oops->{R385} = $oops->{R384}; $oops->{R386} = \$oops->{A500}; $oops->commit; SQL queries require a bunch of joins to link data structures together. Here are some examples. "SELECT Foobar WHERE $Foobar->{xyz} = 'abc'" SELECT object.id FROM object, attribute WHERE object.class = 'Foobar' AND object.id = attribute.id AND object.otype = 'H' AND attribute.pkey = 'xyz' AND attribute.pval = 'abc' AND attribute.ptype = '0' "SELECT Foobar WHERE ${$Foobar->{xyz}} = 'abc'" This example should show why an automatic translator would be a good idea... SELECT ohash.object FROM object AS ohash, attribute AS ahash, object AS oref, attribute AS aref, attribute AS target WHERE ohash.class = 'Foobar' AND ohash.otype = 'H' AND ahash.id = ohash.id AND ahash.pkey = 'xyz' AND ahash.ptype = 'R' AND oref.id = ahash.pval AND oref.otype = 'S' # this is the outer ref AND oref.id = aref.id AND aref.pval = target.pkey # here's the reference indirection AND target.pval = 'abc' AND target.ptype = '0' If you construct a query like these examples that return object id's, then use "$object = $oops->load_object($id)" to load them into memory. I recommend that hand-written queries be read-only as there are additional columns that must be kept consistent. For example, the object table includes a reference count column to handle garbage collection of the persistent data. RUNNING THE REGRESSION TEST SUITE The regression test suite empties and re-creates the persistent store over and over again. To prevent the accidental erasure of production data, all of the tests require a special environment variable to be set $OOPSTEST_DSN. This variable replaces the normal $DBI_DSN or $OOPS_DSN. Correspondingly there is a $OOPSTEST_USER, $OOPSTEST_PASS, and $OOPSTEST_PREFIX. Set these variables to something different than what you use for your production data! Most of the tests take a long time to run and are disabled by default. If you can run the full suite in less than six hours please tell me about your configuration. Beware mysql logging. On Debian unstable, the default configuration for mysql used to log every SQL statement. Running the test suite to completion will generate several gigabytes of log file. Running out of disk space will cause the tests to fail. On DragonflyBSD the default mysql configuration includes making replication master logs. THE COMPETITION There are a number of other modules that make perl objects persistent. Other object persistence solutions either store data in blobs that cannot be querried or they they're a mapping from a relation database into perl data that does not allow arbitrary data to be stored. has an overview of options. EXAMPLE PROGRAMS The following applications use OOPS: Qpsmtpd::Plugin::Quarantine LICENSE Copyright(C) 2004-2006 David Muir Sharnoff Copyright(C) 2012 Google, Inc OOPS may be used, modifified and redistributed on the same terms as Perl itself.