# Automatically built by dist/s_test; may require local editing. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= Cold-boot a 4-site group. The first two sites start quickly and initiate an election. The other two sites don't join the election until the middle of the long full election timeout period. It's important that the number of sites that start immediately be a sub-majority, because that's the case that used to have a bug in it [#18456]. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= backup Test of hotbackup functionality. Do all the of the following tests with and without the -c (checkpoint) option; and with and without the transactional bulk loading optimization. Make sure that -c and -d (data_dir) are not allowed together. (1) Test that plain and simple hotbackup works. (2) Test with -data_dir (-d). (3) Test updating an existing hot backup (-u). (4) Test with absolute path. (5) Test with DB_CONFIG, setting log_dir (-l) and data_dir (-d). (6) DB_CONFIG and update. (7) Repeat hot backup (non-update) with DB_CONFIG, DB_CONFIG (-D) and existing directories. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= bigfile001 Create a database greater than 4 GB in size. Close, verify. Grow the database somewhat. Close, reverify. Lather, rinse, repeat. Since it will not work on all systems, this test is not run by default. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= bigfile002 This one should be faster and not require so much disk space, although it doesn't test as extensively. Create an mpool file with 1K pages. Dirty page 6000000. Sync. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= db_reptest Wrapper to configure and run the db_reptest program. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= dbm Historic DBM interface test. Use the first 1000 entries from the dictionary. Insert each with self as key and data; retrieve each. After all are entered, retrieve all; compare output to original. Then reopen the file, re-retrieve everything. Finally, delete everything. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= dead001 Use two different configurations to test deadlock detection among a variable number of processes. One configuration has the processes deadlocked in a ring. The other has the processes all deadlocked on a single resource. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= dead002 Same test as dead001, but use "detect on every collision" instead of separate deadlock detector. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= dead003 Same test as dead002, but explicitly specify DB_LOCK_OLDEST and DB_LOCK_YOUNGEST. Verify the correct lock was aborted/granted. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= dead006 use timeouts rather than the normal dd algorithm. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= dead007 Tests for locker and txn id wraparound. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= dead008 Run dead001 deadlock test using priorities =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= dead009 Run dead002 deadlock test using priorities =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= dead010 Same test as dead003, except the actual youngest and oldest will have higher priorities. Verify that the oldest/youngest of the lower priority lockers gets killed. Doesn't apply to 2 procs. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= dead011 Test out the minlocks, maxlocks, and minwrites options to the deadlock detector when priorities are used. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= env001 Test of env remove interface (formerly env_remove). =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= env002 Test of DB_LOG_DIR and env name resolution. With an environment path specified using -home, and then again with it specified by the environment variable DB_HOME: 1) Make sure that the set_lg_dir option is respected a) as a relative pathname. b) as an absolute pathname. 2) Make sure that the DB_LOG_DIR db_config argument is respected, again as relative and absolute pathnames. 3) Make sure that if -both- db_config and a file are present, only the file is respected (see doc/env/naming.html). =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= env003 Test DB_TMP_DIR and env name resolution With an environment path specified using -home, and then again with it specified by the environment variable DB_HOME: 1) Make sure that the DB_TMP_DIR config file option is respected a) as a relative pathname. b) as an absolute pathname. 2) Make sure that the -tmp_dir config option is respected, again as relative and absolute pathnames. 3) Make sure that if -both- -tmp_dir and a file are present, only the file is respected (see doc/env/naming.html). =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= env004 Test multiple data directories. Do a bunch of different opens to make sure that the files are detected in different directories. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= env005 Test that using subsystems without initializing them correctly returns an error. Cannot test mpool, because it is assumed in the Tcl code. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= env006 Make sure that all the utilities exist and run. Test that db_load -r options don't blow up. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= env007 Test DB_CONFIG config file options for berkdb env. 1) Make sure command line option is respected 2) Make sure that config file option is respected 3) Make sure that if -both- DB_CONFIG and the set_ method is used, only the file is respected. Then test all known config options. Also test config options on berkdb open. This isn't really env testing, but there's no better place to put it. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= env008 Test environments and subdirectories. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= env009 Test calls to all the various stat functions. We have several sprinkled throughout the test suite, but this will ensure that we run all of them at least once. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= env010 Run recovery in an empty directory, and then make sure we can still create a database in that directory. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= env011 Run with region overwrite flag. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= env012 Test DB_REGISTER. DB_REGISTER will fail on systems without fcntl. If it fails, make sure we got the expected DB_OPNOTSUP return. Then, the real tests: For each test, we start a process that opens an env with -register. 1. Verify that a 2nd process can enter the existing env with -register. 2. Kill the 1st process, and verify that the 2nd process can enter with "-register -recover". 3. Kill the 1st process, and verify that the 2nd process cannot enter with just "-register". 4. While the 1st process is still running, a 2nd process enters with "-register". Kill the 1st process. Verify that a 3rd process can enter with "-register -recover". Verify that the 3rd process, entering, causes process 2 to fail with the message DB_RUNRECOVERY. 5. We had a bug where recovery was always run with -register if there were empty slots in the process registry file. Verify that recovery doesn't automatically run if there is an empty slot. 6. Verify process cannot connect when specifying -failchk and an isalive function has not been declared. 7. Verify that a 2nd process can enter the existing env with -register and -failchk and having specified an isalive function 8. Kill the 1st process, and verify that the 2nd process can enter with "-register -failchk -recover" 9. 2nd process enters with "-register -failchk". Kill the 1st process. 2nd process may get blocked on a mutex held by process one. Verify 3rd process can enter with "-register -recover -failchk". 3rd process should run failchk, clear out open txn/log from process 1. It will enter env without need for any additional recovery. We look for "Freeing log information .." sentence in the log for 3rd process as an indication that failchk ran. If DB_RUNRECOVERY were returned instead it would mean failchk could not recover. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= env013 Test of basic functionality of fileid_reset. Create a database in an env. Copy it to a new file within the same env. Reset the file id and make sure it has changed. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= env014 Make sure that attempts to open an environment with incompatible flags (e.g. replication without transactions) fail with the appropriate messages. A new thread of control joining an env automatically initializes the same subsystems as the original env. Make sure that the attempt to change subsystems when joining an env fails with the appropriate messages. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= env015 Rename the underlying directory of an env, make sure everything still works. Test runs with regular named databases and with in-memory named databases. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= env016 Replication settings and DB_CONFIG Create a DB_CONFIG for various replication settings. Use rep_stat or getter functions to verify they're set correctly. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= env017 Check documented "stat" fields against the fields returned by the "stat" functions. Make sure they match, and that none are missing. These are the stat functions we test: env log_stat env lock_stat env txn_stat env mutex_stat env rep_stat env repmgr_stat env mpool_stat db stat seq stat db compact_stat =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= env018 Test getters when joining an env. When a second handle is opened on an existing env, get_open_flags needs to return the correct flags to the second handle so it knows what sort of environment it's just joined. For several different flags to env_open, open an env. Open a second handle on the same env, get_open_flags and verify the flag is returned. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= env019 Test that stats are correctly set and reported when an env is accessed from a second process. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= env020 Check if the output information for stat_print is expected. These are the stat_print functions we test: env stat_print env lock_stat_print env log_stat_print env mpool_stat_print env mutex_stat_print env rep_stat_print env repmgr_stat_print env txn_stat_print db stat_print seq stat_print =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= env021 Test the operations on a transaction in a CDS environment. These are the operations we test: $txn abort $txn commit $txn id $txn prepare $txn setname name $txn getname $txn discard $txn set_timeout In these operations, we only support the following: $txn id $txn commit =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= fop001.tcl Test two file system operations combined in one transaction. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= fop002.tcl Test file system operations in the presence of bad permissions. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= fop003 Test behavior of create and truncate for compatibility with sendmail. 1. DB_TRUNCATE is not allowed with locking or transactions. 2. Can -create into zero-length existing file. 3. Can -create into non-zero-length existing file if and only if DB_TRUNCATE is specified. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= fop004 Test of DB->rename(). (formerly test075) Test that files can be renamed from one directory to another. Test that files can be renamed using absolute or relative pathnames. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= fop005 Test of DB->remove() Formerly test080. Test use of dbremove with and without envs, with absolute and relative paths, and with subdirectories. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= fop006 Test file system operations in multiple simultaneous transactions. Start one transaction, do a file operation. Start a second transaction, do a file operation. Abort or commit txn1, then abort or commit txn2, and check for appropriate outcome. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= fop007 Test file system operations on named in-memory databases. Combine two ops in one transaction. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= fop008 Test file system operations on named in-memory databases. Combine two ops in one transaction. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= fop009 Test file system operations in child transactions. Combine two ops in one child transaction. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= fop010 Test file system operations in child transactions. Two ops, each in its own child txn. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= fop011 Test file system operations in child transactions. Combine two ops in one child transaction, with in-emory databases. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= fop012 Test file system operations in child transactions. Two ops, each in its own child txn, with in-memory dbs. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= jointest Test duplicate assisted joins. Executes 1, 2, 3 and 4-way joins with differing index orders and selectivity. We'll test 2-way, 3-way, and 4-way joins and figure that if those work, everything else does as well. We'll create test databases called join1.db, join2.db, join3.db, and join4.db. The number on the database describes the duplication -- duplicates are of the form 0, N, 2N, 3N, ... where N is the number of the database. Primary.db is the primary database, and null.db is the database that has no matching duplicates. We should test this on all btrees, all hash, and a combination thereof =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= lock001 Make sure that the basic lock tests work. Do some simple gets and puts for a single locker. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= lock002 Exercise basic multi-process aspects of lock. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= lock003 Exercise multi-process aspects of lock. Generate a bunch of parallel testers that try to randomly obtain locks; make sure that the locks correctly protect corresponding objects. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= lock004 Test locker ids wraping around. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= lock005 Check that page locks are being released properly. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= lock006 Test lock_vec interface. We do all the same things that lock001 does, using lock_vec instead of lock_get and lock_put, plus a few more things like lock-coupling. 1. Get and release one at a time. 2. Release with put_obj (all locks for a given locker/obj). 3. Release with put_all (all locks for a given locker). Regularly check lock_stat to verify all locks have been released. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= log001 Read/write log records. Test with and without fixed-length, in-memory logging, and encryption. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= log002 Tests multiple logs Log truncation LSN comparison and file functionality. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= log003 Verify that log_flush is flushing records correctly. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= log004 Make sure that if we do PREVs on a log, but the beginning of the log has been truncated, we do the right thing. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= log005 Check that log file sizes can change on the fly. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= log006 Test log file auto-remove. Test normal operation. Test a long-lived txn. Test log_archive flags. Test db_archive flags. Test turning on later. Test setting via DB_CONFIG. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= log007 Test of in-memory logging bugs. [#11505] Test db_printlog with in-memory logs. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= log008 Test what happens if a txn_ckp record falls into a different log file than the DBREG_CKP records generated by the same checkpoint. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= log009 Test of logging and getting log file version information. Each time we cross a log file boundary verify we can get the version via the log cursorlag. Do this both forward and backward. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= memp001 Randomly updates pages. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= memp002 Tests multiple processes accessing and modifying the same files. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= memp003 Test reader-only/writer process combinations; we use the access methods for testing. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= memp004 Test that small read-only databases are mapped into memory. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= memp005 Make sure that db pagesize does not interfere with mpool pagesize. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= mut001 Exercise the mutex API. Allocate, lock, unlock, and free a bunch of mutexes. Set basic configuration options and check mutex_stat and the mutex getters for the correct values. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= mut002 Two-process mutex test. Allocate and lock a self-blocking mutex. Start another process. Try to lock the mutex again -- it will block. Unlock the mutex from the other process, and the blocked lock should be obtained. Clean up. Do another test with a "-process-only" mutex. The second process should not be able to unlock the mutex. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= mut003 Try doing mutex operations out of order. Make sure we get appropriate errors. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= plat001 Test of portability of sequences. Create and dump a database containing sequences. Save the dump. This test is used in conjunction with the upgrade tests, which will compare the saved dump to a locally created dump. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= recd001 Per-operation recovery tests for non-duplicate, non-split messages. Makes sure that we exercise redo, undo, and do-nothing condition. Any test that appears with the message (change state) indicates that we've already run the particular test, but we are running it again so that we can change the state of the data base to prepare for the next test (this applies to all other recovery tests as well). These are the most basic recovery tests. We do individual recovery tests for each operation in the access method interface. First we create a file and capture the state of the database (i.e., we copy it. Then we run a transaction containing a single operation. In one test, we abort the transaction and compare the outcome to the original copy of the file. In the second test, we restore the original copy of the database and then run recovery and compare this against the actual database. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= recd002 Split recovery tests. For every known split log message, makes sure that we exercise redo, undo, and do-nothing condition. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= recd003 Duplicate recovery tests. For every known duplicate log message, makes sure that we exercise redo, undo, and do-nothing condition. Test all the duplicate log messages and recovery operations. We make sure that we exercise all possible recovery actions: redo, undo, undo but no fix necessary and redo but no fix necessary. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= recd004 Big key test where big key gets elevated to internal page. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= recd005 Verify reuse of file ids works on catastrophic recovery. Make sure that we can do catastrophic recovery even if we open files using the same log file id. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= recd006 Nested transactions. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= recd007 File create/delete tests. This is a recovery test for create/delete of databases. We have hooks in the database so that we can abort the process at various points and make sure that the transaction doesn't commit. We then need to recover and make sure the file is correctly existing or not, as the case may be. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= recd008 Test deeply nested transactions and many-child transactions. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= recd009 Verify record numbering across split/reverse splits and recovery. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= recd010 Test stability of btree duplicates across btree off-page dup splits and reverse splits and across recovery. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= recd011 Verify that recovery to a specific timestamp works. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= recd012 Test of log file ID management. [#2288] Test recovery handling of file opens and closes. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= recd013 Test of cursor adjustment on child transaction aborts. [#2373] =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= recd014 This is a recovery test for create/delete of queue extents. We then need to recover and make sure the file is correctly existing or not, as the case may be. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= recd015 This is a recovery test for testing lots of prepared txns. This test is to force the use of txn_recover to call with the DB_FIRST flag and then DB_NEXT. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= recd016 Test recovery after checksum error. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= recd017 Test recovery and security. This is basically a watered down version of recd001 just to verify that encrypted environments can be recovered. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= recd018 Test recover of closely interspersed checkpoints and commits. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= recd019 Test txn id wrap-around and recovery. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= recd020 Test creation of intermediate directories -- an undocumented, UNIX-only feature. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= recd021 Test of failed opens in recovery. If a file was deleted through the file system (and not within Berkeley DB), an error message should appear. Test for regular files and subdbs. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= recd022 Test that pages allocated by an aborted subtransaction within an aborted prepared parent transaction are returned to the free list after recovery. This exercises __db_pg_prepare in systems without FTRUNCATE. [#7403] =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= recd023 Test recover of reverse split. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= recd024 Test recovery of streaming partial insert operations. These are operations that do multiple partial puts that append to an existing data item (as long as the data item is on an overflow page). The interesting cases are: * Simple streaming operations * Operations that cause the overflow item to flow onto another page. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= recd025 Basic tests for transaction bulk loading and recovery. In particular, verify that the tricky hot backup protocol works. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep001 Replication rename and forced-upgrade test. Run rep_test in a replicated master environment. Verify that the database on the client is correct. Next, remove the database, close the master, upgrade the client, reopen the master, and make sure the new master can correctly run rep_test and propagate it in the other direction. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep002 Basic replication election test. Run a modified version of test001 in a replicated master environment; hold an election among a group of clients to make sure they select a proper master from amongst themselves, in various scenarios. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep003 Repeated shutdown/restart replication test Run a quick put test in a replicated master environment; start up, shut down, and restart client processes, with and without recovery. To ensure that environment state is transient, use DB_PRIVATE. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep005 Replication election test with error handling. Run rep_test in a replicated master environment; hold an election among a group of clients to make sure they select a proper master from amongst themselves, forcing errors at various locations in the election path. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep006 Replication and non-rep env handles. Run a modified version of test001 in a replicated master environment; verify that the database on the client is correct. Next, create a non-rep env handle to the master env. Attempt to open the database r/w to force error. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep007 Replication and bad LSNs Run rep_test in a replicated master env. Close the client. Make additional changes to master. Close the master. Open the client as the new master. Make several different changes. Open the old master as the client. Verify periodically that contents are correct. This test is not appropriate for named in-memory db testing because the databases are lost when both envs are closed. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep008 Replication, back up and synchronizing Run a modified version of test001 in a replicated master environment. Close master and client. Copy the master log to the client. Clean the master. Reopen the master and client. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep009 Replication and DUPMASTERs Run test001 in a replicated environment. Declare one of the clients to also be a master. Close a client, clean it and then declare it a 2nd master. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep010 Replication and ISPERM With consecutive message processing, make sure every DB_REP_PERMANENT is responded to with an ISPERM when processed. With gaps in the processing, make sure every DB_REP_PERMANENT is responded to with an ISPERM or a NOTPERM. Verify in both cases that the LSN returned with ISPERM is found in the log. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep011 Replication: test open handle across an upgrade. Open and close test database in master environment. Update the client. Check client, and leave the handle to the client open as we close the masterenv and upgrade the client to master. Reopen the old master as client and catch up. Test that we can still do a put to the handle we created on the master while it was still a client, and then make sure that the change can be propagated back to the new client. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep012 Replication and dead DB handles. Run a modified version of test001 in a replicated master env. Run in replicated environment with secondary indices too. Make additional changes to master, but not to the client. Downgrade the master and upgrade the client with open db handles. Verify that the roll back on clients gives dead db handles. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep013 Replication and swapping master/clients with open dbs. Run a modified version of test001 in a replicated master env. Make additional changes to master, but not to the client. Swap master and client. Verify that the roll back on clients gives dead db handles. Rerun the test, turning on client-to-client synchronization. Swap and verify several times. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep014 Replication and multiple replication handles. Test multiple client handles, opening and closing to make sure we get the right openfiles. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep015 Locking across multiple pages with replication. Open master and client with small pagesize and generate more than one page and generate off-page dups on the first page (second key) and last page (next-to-last key). Within a single transaction, for each database, open 2 cursors and delete the first and last entries (this exercises locks on regular pages). Intermittently update client during the process. Within a single transaction, for each database, open 2 cursors. Walk to the off-page dups and delete one from each end (this exercises locks on off-page dups). Intermittently update client. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep016 Replication election test with varying required nvotes. Run a modified version of test001 in a replicated master environment; hold an election among a group of clients to make sure they select the master with varying required participants. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep017 Concurrency with checkpoints. Verify that we achieve concurrency in the presence of checkpoints. Here are the checks that we wish to make: While dbenv1 is handling the checkpoint record: Subsequent in-order log records are accepted. Accepted PERM log records get NOTPERM A subsequent checkpoint gets NOTPERM After checkpoint completes, next txn returns PERM =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep018 Replication with dbremove. Verify that the attempt to remove a database file on the master hangs while another process holds a handle on the client. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep019 Replication and multiple clients at same LSN. Have several clients at the same LSN. Run recovery at different times. Declare a client master and after sync-up verify all client logs are identical. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep020 Replication elections - test election generation numbers. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep021 Replication and multiple environments. Run similar tests in separate environments, making sure that some data overlaps. Then, "move" one client env from one replication group to another and make sure that we do not get divergent logs. We either match the first record and end up with identical logs or we get an error. Verify all client logs are identical if successful. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep022 Replication elections - test election generation numbers during simulated network partition. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep023 Replication using two master handles. Open two handles on one master env. Create two databases, one through each master handle. Process all messages through the first master handle. Make sure changes made through both handles are picked up properly. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep024 Replication page allocation / verify test Start a master (site 1) and a client (site 2). Master closes (simulating a crash). Site 2 becomes the master and site 1 comes back up as a client. Verify database. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep025 Test of DB_REP_JOIN_FAILURE. One master, one client. Generate several log files. Remove old master log files. Delete client files and restart client. Put one more record to the master. At the next processing of messages, the client should get JOIN_FAILURE. Recover with a hot failover. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep026 Replication elections - simulate a crash after sending a vote. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep027 Replication and secondary indexes. Set up a secondary index on the master and make sure it can be accessed from the client. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep028 Replication and non-rep env handles. (Also see rep006.) Open second non-rep env on client, and create a db through this handle. Open the db on master and put some data. Check whether the non-rep handle keeps working. Also check if opening the client database in the non-rep env writes log records. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep029 Test of internal initialization. One master, one client. Generate several log files. Remove old master log files. Delete client files and restart client. Put one more record to the master. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep030 Test of internal initialization multiple files and pagesizes. Hold some databases open on master. One master, one client using a data_dir for internal init. Generate several log files. Remove old master log files. Delete client files and restart client. Put one more record to the master. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep031 Test of internal initialization and blocked operations. One master, one client. Put one more record to the master. Test that internal initialization blocks: log_archive, rename, remove, fileid_reset, lsn_reset. Sleep 30+ seconds. Test that blocked operations are now unblocked. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep032 Test of log gap processing. One master, one client. Run rep_test. Run rep_test without sending messages to client. Make sure client missing the messages catches up properly. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep033 Test of internal initialization with rename and remove of dbs. One master, one client. Generate several databases. Replicate to client. Do some renames and removes, both before and after closing the client. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep034 Test of STARTUPDONE notification. STARTUPDONE can now be recognized without the need for new "live" log records from the master (under favorable conditions). The response to the ALL_REQ at the end of synchronization includes an end-of-log marker that now triggers it. However, the message containing that end marker could get lost, so live log records still serve as a back-up mechanism. The end marker may also be set under c2c sync, but only if the serving client has itself achieved STARTUPDONE. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep035 Test sync-up recovery in replication. We need to fork off 3 child tclsh processes to operate on Site 3's (client always) home directory: Process 1 continually calls lock_detect. Process 2 continually calls txn_checkpoint. Process 3 continually calls memp_trickle. Process 4 continually calls log_archive. Sites 1 and 2 will continually swap being master (forcing site 3 to continually run sync-up recovery) New master performs 1 operation, replicates and downgrades. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep036 Multiple master processes writing to the database. One process handles all message processing. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep037 Test of internal initialization and page throttling. One master, one client, force page throttling. Generate several log files. Remove old master log files. Delete client files and restart client. Put one more record to the master. Verify page throttling occurred. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep038 Test of internal initialization and ongoing master updates. One master, one client. Generate several log files. Remove old master log files. Delete client files and restart client. Put more records on master while initialization is in progress. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep039 Test of interrupted internal initialization. The interruption is due to a changed master, or the client crashing, or both. One master, two clients. Generate several log files. Remove old master log files. Restart client, optionally having "cleaned" client env dir. Either way, this has the effect of forcing an internal init. Interrupt the internal init. Vary the number of times we process messages to make sure the interruption occurs at varying stages of the first internal initialization. Run for btree and queue only because of the number of permutations. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep040 Test of racing rep_start and transactions. One master, one client. Have master in the middle of a transaction. Call rep_start to make master a client. Commit the transaction. Call rep_start to make master the master again. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep041 Turn replication on and off at run-time. Start a master with replication OFF (noop transport function). Run rep_test to advance log files and archive. Start up client; change master to working transport function. Now replication is ON. Do more ops, make sure client is up to date. Close client, turn replication OFF on master, do more ops. Repeat from point A. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep042 Concurrency with updates. Verify racing role changes and updates don't result in pages with LSN 0,1. Set up an environment that is master. Spawn child process that does a delete, but using the $env check so that it sleeps in the middle of the call. Master downgrades and then sleeps as a client so that child will run. Verify child does not succeed (should get read-only error) due to role change in the middle of its call. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep043 Constant writes during upgrade/downgrade. Three envs take turns being master. Each env has a child process which does writes all the time. They will succeed when that env is master and fail when it is not. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep044 Test rollbacks with open file ids. We have one master with two handles and one client. Each time through the main loop, we open a db, write to the db, and close the db. Each one of these actions is propagated to the client, or a roll back is forced by swapping masters. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep045 Replication with versions. Mimic an application where a database is set up in the background and then put into a replication group for use. The "version database" identifies the current live version, the database against which queries are made. For example, the version database might say the current version is 3, and queries would then be sent to db.3. Version 4 is prepared for use while version 3 is in use. When version 4 is complete, the version database is updated to point to version 4 so queries can be directed there. This test has a master and two clients. One client swaps roles with the master, and the other client runs constantly in another process. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep046 Replication and basic bulk transfer. Set bulk transfer replication option. Run long txns on master and then commit. Process on client and verify contents. Run a very long txn so that logging must send the log. Process and verify on client. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep047 Replication and log gap bulk transfers. Set bulk transfer replication option. Run test. Start a new client (to test ALL_REQ and bulk). Run small test again. Clear messages for 1 client. Run small test again to test LOG_REQ gap processing and bulk. Process and verify on clients. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep048 Replication and log gap bulk transfers. Have two master env handles. Turn bulk on in one (turns it on for both). Turn it off in the other. While toggling, send log records from both handles. Process message and verify master and client match. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep049 Replication and delay syncing clients - basic test. Open and start up a master and two clients. Turn on delay sync in the delayed client. Change master, add data and process messages. Verify delayed client does not match. Make additional changes and update the delayted client. Verify all match. Add in a fresh delayed client to test delay of ALL_REQ. Process startup messages and verify freshc client has no database. Sync and verify fresh client matches. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep050 Replication and delay syncing clients - change master test. Open and start up master and 4 clients. Turn on delay for 3 clients. Switch masters, add data and verify delayed clients are out of date. Make additional changes to master. And change masters again. Sync/update delayed client and verify. The 4th client is a brand new delayed client added in to test the non-verify path. Then test two different things: 1. Swap master again while clients are still delayed. 2. Swap master again while sync is proceeding for one client. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep051 Test of compaction with replication. Run rep_test in a replicated master environment. Delete a large number of entries and compact with -freespace. Propagate the changes to the client and make sure client and master match. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep052 Test of replication with NOWAIT. One master, one client. After initializing everything normally, close client and let the master get ahead -- far enough that the master no longer has the client's last log file. Reopen the client and turn on NOWAIT. Process a few messages to get the client into recovery mode, and verify that lockout occurs on a txn API call (txn_begin) and an env API call. Process all the messages and verify that lockout is over. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep053 Replication and basic client-to-client synchronization. Open and start up master and 1 client. Start up a second client later and verify it sync'ed from the original client, not the master. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep054 Test of internal initialization where a far-behind client takes over as master. One master, two clients. Run rep_test and process. Close client 1. Run rep_test, opening new databases, and processing messages. Archive as we go so that log files get removed. Close master and reopen client 1 as master. Process messages. Verify that new master and client are in sync. Run rep_test again, adding data to one of the new named databases. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep055 Test of internal initialization and log archiving. One master, one client. Generate several log files. Remove old master log files and generate several more. Get list of archivable files from db_archive and restart client. As client is in the middle of internal init, remove the log files returned earlier by db_archive. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep058 Replication with early databases Mimic an application where they create a database before calling rep_start, thus writing log records on a client before it is a client. Verify we cannot join repl group. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep060 Test of normally running clients and internal initialization. Have a client running normally, but slow/far behind the master. Then the master checkpoints and archives, causing the client to suddenly be thrown into internal init. This test tests that we clean up the old files/pages in mpool and dbreg. Also test same thing but the app holding an open dbp as well. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep061 Test of internal initialization multiple files and pagesizes with page gaps. One master, one client. Generate several log files. Remove old master log files. Delete client files and restart client. Put one more record to the master. Force some page messages to get dropped. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep062 Test of internal initialization where client has a different kind of database than the master. Create a master of one type, and let the client catch up. Close the client. Remove the database on the master, and create a new database of the same name but a different type. Run the master ahead far enough that internal initialization will be required on the reopen of the client. Reopen the client and verify. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep063 Replication election test with simulated different versions for each site. This tests that old sites with real priority trump ELECTABLE sites with zero priority even with greater LSNs. There is a special case in the code for testing that if the priority is <= 10, we simulate mixed versions for elections. Run a rep_test in a replicated master environment and close; hold an election among a group of clients to make sure they select the master with varying LSNs and priorities. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep064 Replication rename and forced-upgrade test. The test verifies that the client correctly (internally) closes files when upgrading to master. It does this by having the master have a database open, then crashing. The client upgrades to master, and attempts to remove the open database. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep065 Tests replication running with different versions. This capability is introduced with 4.5. Start a replication group of 1 master and N sites, all running some historical version greater than or equal to 4.4. Take down a client and bring it up again running current. Run some upgrades, make sure everything works. Each site runs the tcllib of its own version, but uses the current tcl code (e.g. test.tcl). =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep066 Replication and dead log handles. Run rep_test on master and a client. Simulate client crashes (master continues) until log 2. Open 2nd master env handle and put something in log and flush. Downgrade master, restart client as master. Run rep_test on newmaster until log 2. New master writes log records, newclient processes records and 2nd newclient env handle calls log_flush. New master commits, newclient processes and should succeed. Make sure 2nd handle detects the old log handle and doesn't write to a stale handle (if it does, the processing of the commit will fail). =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep067 Full election timeout test. Verify that elections use a separate "full election timeout" (if such configuration is in use) instead of the normal timeout, when the replication group is "cold-booted" (all sites starting with recovery). =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep068 Verify replication of dbreg operations does not hang clients. In a simple replication group, create a database with very little data. With DB_TXN_NOSYNC the database can be created at the client even though the log is not flushed. If we crash and restart, the application of the log starts over again, even though the database is still there. The application can open the database before replication tries to re-apply the create. This causes a hang as replication waits to be able to get a handle lock. Run for btree only because access method shouldn't matter. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep069 Test of internal initialization and elections. If a client is in a recovery mode of any kind, it participates in elections at priority 0 so it can never be elected master. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep070 Test of startup_done condition with idle master. Join a client to an existing master, and verify that the client detects startup_done even if the master does not execute any new transactions. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep071 Test of multiple simultaneous client env handles and upgrading/downgrading. Tests use of temp db handle internally. Open a master and 2 handles to the same client env. Run rep_test. Close master and upgrade client to master using one env handle. Run rep_test again, and then downgrade back to client. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep072 Verify that internal init does not leak resources from the locking subsystem. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep073 Test of allowing clients to create and update their own scratch databases within the environment. Doing so requires the use use of the DB_TXN_NOT_DURABLE flag for those databases. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep074 Verify replication withstands send errors processing requests. Run for btree only because access method shouldn't matter. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep075 Replication and prepared transactions. Test having outstanding prepared transactions and simulating crashing or upgrading or downgrading sites. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep076 Replication elections - what happens if elected client does not become master? Set up a master and 3 clients. Take down master, run election. The elected client will ignore the fact that it's been elected, so we still have 2 clients. Run another election, a regular election that allows the winner to become master, and make sure it goes okay. We do this both for the client that ignored its election and for the other client. This simulates what would happen if, say, we had a temporary network partition and lost the winner. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep077 Replication, recovery and applying log records immediately. Master and 1 client. Start up both sites. Close client and run rep_test on the master so that the log record is the same LSN the client would be expecting. Reopen client with recovery and verify the client does not try to apply that "expected" record before it synchronizes with the master. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep078 Replication and basic lease test. Set leases on master and 2 clients. Do a lease operation and process to all clients. Read with lease on master. Do another lease operation and don't process on any client. Try to read with on the master and verify it fails. Process the messages to the clients and retry the read. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep079 Replication leases and invalid usage. Open a client without leases. Attempt to set leases after rep_start. Attempt to declare as master without election. Run an election with an nsites parameter value. Elect a master with leases. Put some data and send to clients. Cleanly shutdown master env. Restart without recovery and verify leases are expired and refreshed. Add a new client without leases to a group using leases. Test errors if we cannot get leases before/after txn_commit. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep080 AUTOINIT off with empty client logs. Verify that a fresh client trying to join the group for the first time observes the setting of DELAY_SYNC and !AUTOINIT properly. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep081 Test of internal initialization and missing database files. One master, one client, two databases. Generate several log files. Remove old master log files. Start up client. Remove or replace one master database file while client initialization is in progress, make sure other master database can keep processing. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep082 Sending replication requests to correct master site. Regression test for a bug [#16592] where a client could send an UPDATE_REQ to another client instead of the master. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep083 Replication clients must never send VERIFY_FAIL to a c2c request. Regression test for a bug [#16592] where a client could send a VERIFY_FAIL to another client, which is illegal. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep084 Abbreviated internal init for named in-memory databases (NIMDBs). =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep085 Skipping unnecessary abbreviated internal init. Make sure that once we've materialized NIMDBs, we don't bother trying to do it again on subsequent sync without recovery. Make sure we do probe for the need to materialize NIMDBs, but don't do any internal init at all if there are no NIMDBs. Note that in order to do this test we don't even need any NIMDBs. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep086 Interrupted abbreviated internal init. Make sure we cleanly remove partially loaded named in-memory databases (NIMDBs). =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep087 Abbreviated internal init with open file handles. Client has open handle to an on-disk DB when abbreviated internal init starts. Make sure we lock out access, and make sure it ends up as HANDLE_DEAD. Also, make sure that if there are no NIMDBs, that we *don't* get HANDLE_DEAD. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep088 Replication roll-back preserves checkpoint. Create a situation where a client has to roll back its log, discarding some existing transactions, in order to sync with a new master. 1. When the client still has its entire log file history, all the way back to log file #1, it's OK if the roll-back discards any/all checkpoints. 2. When old log files have been archived, if the roll-back would remove all existing checkpoints it must be forbidden. The log must always have a checkpoint (or all files back through #1). The client must do internal init or return JOIN_FAILURE. 3. (the normal case) Old log files archived, and a checkpoint still exists in the portion of the log which will remain after the roll-back: no internal-init/JOIN_FAILURE necessary. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep089 Test of proper clean-up of mpool during interrupted internal init. Have a client in the middle of internal init when a new master generation comes along, forcing the client to interrupt the internal init, including doing the clean-up. The client is in the middle of retrieving database pages, so that we are forced to clean up mpool. (Regression test for bug 17121) =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep090 Test of AUTO_REMOVE on both master and client sites. One master, one client. Set AUTO_REMOVE on the client env. Generate several log files. Verify the client has properly removed the log files. Turn on AUTO_REMOVE on the master and generate more log files. Confirm both envs have the same log files. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep091 Read-your-writes consistency. Write transactions at the master, and then call the txn_applied() method to see whether the client has received and applied them yet. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep092 Read-your-writes consistency. Test events in one thread (process) waking up another sleeping thread, before a timeout expires. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep093 Egen changes during election. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep094 Full election with less than majority initially connected. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep095 Test of internal initialization use of shared region memory. One master, one client. Create a gap that requires internal initialization. Start the internal initialization in this parent process and complete it in a separate child process. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep096 Replication and db_replicate utility. Create a master and client environment. Open them. Start a db_replicate process on each. Create a database on the master and write some data. Then verify contents. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep097 Replication and lease data durability test. Set leases on master and 2 clients. Have the original master go down and a client take over. Have the old master rejoin as client, but go down again. The other two sites do one txn, while the original master's LSN extends beyond due to running recovery. Original Master rejoins while new master fails. Make sure remaining original site is elected, with the smaller LSN, but with txn data. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep098 Test of internal initialization and page deallocation. Use one master, one client. Generate several log files. Remove old master log files. Start a client. After client gets UPDATE file information, delete entries to remove pages in the database. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep099 Test of multiple data dirs and databases, both active replication and internal init. One master, two clients using several data_dirs. Create databases in different data_dirs. Replicate to client. Add 2nd client later to require it to catch up via internal init. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep100 Checkpoints and unresolved txns =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep101 Test of exclusive access to databases and HA. Confirm basic functionality with master and client. Confirm excl access after checkpoint and archive. Confirm internal init with open excl databases. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rep102 Test of exclusive access to databases and HA. Master creates a database, writes some txns, and closes the db. Client opens the database (in child process). Master opens database exclusively. Client tries to read database and gets HANDLE_DEAD. Closes db. Client tries to reopen database and is blocked. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr001 Basic repmgr test. Run all mix-and-match combinations of the basic_repmgr_test. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr002 Basic repmgr test. Run all combinations of the basic_repmgr_election_test. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr003 Basic repmgr init test. Run all combinations of the basic_repmgr_init_test. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr007 Basic repmgr client shutdown/restart test. Start an appointed master site and two clients. Shut down and restart each client, processing transactions after each restart. Verify all expected transactions are replicated. Run for btree only because access method shouldn't matter. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr009 repmgr API error test. Try a variety of repmgr calls that result in errors. Also try combinations of repmgr and base replication API calls that result in errors. Run for btree only because access method shouldn't matter. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr010 Acknowledgement policy and timeout test. Verify that "quorum" acknowledgement policy succeeds with fewer than nsites running. Verify that "all" acknowledgement policy results in ack failures with fewer than nsites running. Run for btree only because access method shouldn't matter. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr011 repmgr two site strict majority test. Start an appointed master and one client with 2 site strict majority set. Shut down the master site, wait and verify that the client site was not elected master. Start up master site and verify that transactions are processed as expected. Run for btree only because access method shouldn't matter. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr012 repmgr heartbeat test. Start an appointed master and one client site. Set heartbeat send and monitor values and process some transactions. Stop sending heartbeats from master and verify that client sees a dropped connection. Run for btree only because access method shouldn't matter. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr013 Site list test. Configure a master and two clients where one client is a peer of the other and verify resulting site lists. Run for btree only because access method shouldn't matter. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr017 repmgr in-memory cache overflow test. Start an appointed master site and one client, putting databases, environment regions, logs and replication files in-memory. Set very small cachesize and run enough transactions to overflow cache. Shut down and restart master and client, giving master a larger cache. Run and verify a small number of transactions. Run for btree only because access method shouldn't matter. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr018 Check repmgr stats. Start an appointed master and one client. Shut down the client, run some transactions at the master and verify that there are acknowledgement failures and one dropped connection. Shut down and restart client again and verify that there are two dropped connections. Run for btree only because access method shouldn't matter. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr023 Test of JOIN_FAILURE event for repmgr applications. Run for btree only because access method shouldn't matter. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr024 Test of group-wide log archiving awareness. Verify that log archiving will use the ack from the clients in its decisions about what log files are allowed to be archived. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr025 repmgr heartbeat rerequest test. Start an appointed master site and one client. Use a test hook to inhibit PAGE_REQ processing at the master (i.e., "lose" some messages). Start a second client that gets stuck in internal init. Wait long enough to rely on the heartbeat rerequest to request the missing pages, rescind the test hook and verify that all data appears on both clients. Run for btree only because access method shouldn't matter. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr026 Test of "full election" timeouts. 1. Cold boot with all sites present. 2. Cold boot with some sites missing. 3. Partial-participation election with one client having seen a master, but another just starting up fresh. 4. Partial participation, with all participants already having seen a master. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr027 Test of "full election" timeouts, where a client starts up and joins the group during the middle of an election. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr028 Repmgr allows applications to choose master explicitly, instead of relying on elections. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr029 Test repmgr group membership: create, join, re-join and remove from repmgr group and observe changes in group membership database. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr030 repmgr multiple client-to-client peer test. Start an appointed master and three clients. The third client configures the other two clients as peers and delays client sync. Add some data and confirm that the third client uses first client as a peer. Close the master so that the first client now becomes the master. Add some more data and confirm that the third client now uses the second client as a peer. Run for btree only because access method shouldn't matter. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr032 The (undocumented) AUTOROLLBACK config feature. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr033 Under quorum policy, if the number of peers in the group is less than a majority, that's still OK. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr034 Repmgr site removal and restart. Start two repmgr sites, master and client1. The client1 removes itself and restarts, sometimes more than once. Start a third repmgr site client2 and make sure it can remove client1 from the group. This test must use heartbeats to ensure that a client that has been removed and restarted without recovery can sync up. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr100 Basic test of repmgr's multi-process master support. Set up a simple 2-site group, create data and replicate it. Add a second process at the master and have it write some updates. It does not explicitly start repmgr (nor do any replication configuration, for that matter). Its first update triggers initiation of connections, and so it doesn't get to the client without a log request. But later updates should go directly. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr101 Repmgr support for multi-process master. Start two processes at the master. Add a client site (not previously known to the master processes), and make sure both master processes connect to it. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr102 Ensuring exactly one listener process. Start a repmgr process with a listener. Start a second process, and see that it does not become the listener. Shut down the first process (gracefully). Now a second process should become listener. Kill the listener process abruptly. Running failchk should show that recovery is necessary. Run recovery and start a clean listener. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr105 Repmgr recognition of peer setting, across processes. Set up a master and two clients, synchronized with some data. Add a new client, configured to use c2c sync with one of the original clients. Check stats to make sure the correct c2c peer was used. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr106 Simple smoke test for repmgr elections with multi-process envs. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr107 Repmgr combined with replication-unaware process at master. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr108 Subordinate connections and processes should not trigger elections. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr109 Test repmgr's internal juggling of peer EID's. Set up master and 2 clients, A and B. Add a third client (C), with two processes. The first process will be configured to know about A. The second process will know about B, and set that as peer, but when it joins the env site B will have to be shuffled into a later position in the list, because A is already first. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr110 Multi-process repmgr start-up policies. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr111 Multi-process repmgr with env open before set local site. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= repmgr112 Multi-process repmgr ack policies. Subordinate processes sending live log records must observe the ack policy set by the main process. Also, a policy change made by a subordinate process should be observed by all processes. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rsrc001 Recno backing file test. Try different patterns of adding records and making sure that the corresponding file matches. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rsrc002 Recno backing file test #2: test of set_re_delim. Specify a backing file with colon-delimited records, and make sure they are correctly interpreted. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rsrc003 Recno backing file test. Try different patterns of adding records and making sure that the corresponding file matches. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= rsrc004 Recno backing file test for EOF-terminated records. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= scr### The scr### directories are shell scripts that test a variety of things, including things about the distribution itself. These tests won't run on most systems, so don't even try to run them. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= sdb001 Tests mixing db and subdb operations Tests mixing db and subdb operations Create a db, add data, try to create a subdb. Test naming db and subdb with a leading - for correct parsing Existence check -- test use of -excl with subdbs Test non-subdb and subdb operations Test naming (filenames begin with -) Test existence (cannot create subdb of same name with -excl) =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= sdb002 Tests basic subdb functionality Small keys, small data Put/get per key Dump file Close, reopen Dump file Use the first 10,000 entries from the dictionary. Insert each with self as key and data; retrieve each. After all are entered, retrieve all; compare output to original. Close file, reopen, do retrieve and re-verify. Then repeat using an environment. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= sdb003 Tests many subdbs Creates many subdbs and puts a small amount of data in each (many defaults to 1000) Use the first 1000 entries from the dictionary as subdbnames. Insert each with entry as name of subdatabase and a partial list as key/data. After all are entered, retrieve all; compare output to original. Close file, reopen, do retrieve and re-verify. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= sdb004 Tests large subdb names subdb name = filecontents, key = filename, data = filecontents Put/get per key Dump file Dump subdbs, verify data and subdb name match Create 1 db with many large subdbs. Use the contents as subdb names. Take the source files and dbtest executable and enter their names as the key with their contents as data. After all are entered, retrieve all; compare output to original. Close file, reopen, do retrieve and re-verify. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= sdb005 Tests cursor operations in subdbs Put/get per key Verify cursor operations work within subdb Verify cursor operations do not work across subdbs =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= sdb006 Tests intra-subdb join We'll test 2-way, 3-way, and 4-way joins and figure that if those work, everything else does as well. We'll create test databases called sub1.db, sub2.db, sub3.db, and sub4.db. The number on the database describes the duplication -- duplicates are of the form 0, N, 2N, 3N, ... where N is the number of the database. Primary.db is the primary database, and sub0.db is the database that has no matching duplicates. All of these are within a single database. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= sdb007 Tests page size difference errors between subdbs. If the physical file already exists, we ignore pagesize specifications on any subsequent -creates. 1. Create/open a subdb with system default page size. Create/open a second subdb specifying a different page size. The create should succeed, but the pagesize of the new db will be the system default page size. 2. Create/open a subdb with a specified, non-default page size. Create/open a second subdb specifying a different page size. The create should succeed, but the pagesize of the new db will be the specified page size from the first create. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= sdb008 Tests explicit setting of lorders for subdatabases -- the lorder should be ignored. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= sdb009 Test DB->rename() method for subdbs =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= sdb010 Test DB->remove() method and DB->truncate() for subdbs =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= sdb011 Test deleting Subdbs with overflow pages Create 1 db with many large subdbs. Test subdatabases with overflow pages. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= sdb012 Test subdbs with locking and transactions Tests creating and removing subdbs while handles are open works correctly, and in the face of txns. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= sdb013 Tests in-memory subdatabases. Create an in-memory subdb. Test for persistence after overflowing the cache. Test for conflicts when we have two in-memory files. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= sdb014 Tests mixing in-memory named and in-memory unnamed dbs. Create a regular in-memory db, add data. Create a named in-memory db. Try to create the same named in-memory db again (should fail). Try to create a different named in-memory db (should succeed). =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= sdb015 Tests basic in-memory named database functionality Small keys, small data Put/get per key Use the first 10,000 entries from the dictionary. Insert each with self as key and data; retrieve each. After all are entered, retrieve all; compare output to original. Close file, reopen, do retrieve and re-verify. Then repeat using an environment. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= sdb016 Creates many in-memory named dbs and puts a small amount of data in each (many defaults to 100) Use the first 100 entries from the dictionary as names. Insert each with entry as name of subdatabase and a partial list as key/data. After all are entered, retrieve all; compare output to original. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= sdb017 Test DB->rename() for in-memory named databases. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= sdb018 Tests join of in-memory named databases. We'll test 2-way, 3-way, and 4-way joins and figure that if those work, everything else does as well. We'll create test databases called sub1.db, sub2.db, sub3.db, and sub4.db. The number on the database describes the duplication -- duplicates are of the form 0, N, 2N, 3N, ... where N is the number of the database. Primary.db is the primary database, and sub0.db is the database that has no matching duplicates. All of these are within a single database. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= sdb019 Tests in-memory subdatabases. Create an in-memory subdb. Test for persistence after overflowing the cache. Test for conflicts when we have two in-memory files. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= sdb020 Tests in-memory subdatabases. Create an in-memory subdb with one page size. Close, and open with a different page size: should succeed, but the underlying page size is not changed. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= sdbtest001 Tests multiple access methods in one subdb Open several subdbs, each with a different access method Small keys, small data Put/get per key per subdb Dump file, verify per subdb Close, reopen per subdb Dump file, verify per subdb Make several subdb's of different access methods all in one DB. Rotate methods and repeat [#762]. Use the first 10,000 entries from the dictionary. Insert each with self as key and data; retrieve each. After all are entered, retrieve all; compare output to original. Close file, reopen, do retrieve and re-verify. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= sdbtest002 Tests multiple access methods in one subdb access by multiple processes. Open several subdbs, each with a different access method Small keys, small data Put/get per key per subdb Fork off several child procs to each delete selected data from their subdb and then exit Dump file, verify contents of each subdb is correct Close, reopen per subdb Dump file, verify per subdb Make several subdb's of different access methods all in one DB. Fork of some child procs to each manipulate one subdb and when they are finished, verify the contents of the databases. Use the first 10,000 entries from the dictionary. Insert each with self as key and data; retrieve each. After all are entered, retrieve all; compare output to original. Close file, reopen, do retrieve and re-verify. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= sec001 Test of security interface =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= sec002 Test of security interface and catching errors in the face of attackers overwriting parts of existing files. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= si001 Secondary index put/delete with lorder test Put data in primary db and check that pget on secondary index finds the right entries. Alter the primary in the following ways, checking for correct data each time: Overwrite data in primary database. Delete half of entries through primary. Delete half of remaining entries through secondary. Append data (for record-based primaries only). =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= si002 Basic cursor-based secondary index put/delete test Cursor put data in primary db and check that pget on secondary index finds the right entries. Open and use a second cursor to exercise the cursor comparison API on secondaries. Overwrite while walking primary, check pget again. Overwrite while walking secondary (use c_pget), check pget again. Cursor delete half of entries through primary, check. Cursor delete half of remainder through secondary, check. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= si003 si001 with secondaries created and closed mid-test Basic secondary index put/delete test with secondaries created mid-test. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= si004 si002 with secondaries created and closed mid-test Basic cursor-based secondary index put/delete test, with secondaries created mid-test. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= si005 Basic secondary index put/delete test with transactions =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= si006 Test -immutable_key interface. DB_IMMUTABLE_KEY is an optimization to be used when a secondary key will not be changed. It does not prevent a deliberate change to the secondary key, it just does not propagate that change when it is made to the primary. This test verifies that a change to the primary is propagated to the secondary or not as specified by -immutable_key. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= si007 Secondary index put/delete with lorder test This test is the same as si001 with the exception that we create and populate the primary and THEN create the secondaries and associate them with -create. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= si008 Secondary index put/delete with lorder test This test is the same as si001 except that we create the secondaries with different byte orders: one native, one swapped. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= sijointest: Secondary index and join test. This used to be si005.tcl. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= sql001 Test db_replicate using a simple SQL app. Start db_replicate on master side and client side, and do various operations using dbsql on master side. After every operation, we will check the records on both sides, to make sure we get same results from both sides. Also try an insert operation on client side; it should fail. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test001 Small keys/data Put/get per key Dump file Close, reopen Dump file Use the first 10,000 entries from the dictionary. Insert each with self as key and data; retrieve each. After all are entered, retrieve all; compare output to original. Close file, reopen, do retrieve and re-verify. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test002 Small keys/medium data Put/get per key Dump file Close, reopen Dump file Use the first 10,000 entries from the dictionary. Insert each with self as key and a fixed, medium length data string; retrieve each. After all are entered, retrieve all; compare output to original. Close file, reopen, do retrieve and re-verify. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test003 Small keys/large data Put/get per key Dump file Close, reopen Dump file Take the source files and dbtest executable and enter their names as the key with their contents as data. After all are entered, retrieve all; compare output to original. Close file, reopen, do retrieve and re-verify. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test004 Small keys/medium data Put/get per key Sequential (cursor) get/delete Check that cursor operations work. Create a database. Read through the database sequentially using cursors and delete each element. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test005 Small keys/medium data Put/get per key Close, reopen Sequential (cursor) get/delete Check that cursor operations work. Create a database; close it and reopen it. Then read through the database sequentially using cursors and delete each element. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test006 Small keys/medium data Put/get per key Keyed delete and verify Keyed delete test. Create database. Go through database, deleting all entries by key. Then do the same for unsorted and sorted dups. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test007 Small keys/medium data Put/get per key Close, reopen Keyed delete Check that delete operations work. Create a database; close database and reopen it. Then issues delete by key for each entry. (Test006 plus reopen) =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test008 Small keys/large data Put/get per key Loop through keys by steps (which change) ... delete each key at step ... add each key back ... change step Confirm that overflow pages are getting reused Take the source files and dbtest executable and enter their names as the key with their contents as data. After all are entered, begin looping through the entries; deleting some pairs and then readding them. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test009 Small keys/large data Same as test008; close and reopen database Check that we reuse overflow pages. Create database with lots of big key/data pairs. Go through and delete and add keys back randomly. Then close the DB and make sure that we have everything we think we should. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test010 Duplicate test Small key/data pairs. Use the first 10,000 entries from the dictionary. Insert each with self as key and data; add duplicate records for each. After all are entered, retrieve all; verify output. Close file, reopen, do retrieve and re-verify. This does not work for recno =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test011 Duplicate test Small key/data pairs. Test DB_KEYFIRST, DB_KEYLAST, DB_BEFORE and DB_AFTER. To test off-page duplicates, run with small pagesize. Use the first 10,000 entries from the dictionary. Insert each with self as key and data; add duplicate records for each. Then do some key_first/key_last add_before, add_after operations. This does not work for recno To test if dups work when they fall off the main page, run this with a very tiny page size. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test012 Large keys/small data Same as test003 except use big keys (source files and executables) and small data (the file/executable names). Take the source files and dbtest executable and enter their contents as the key with their names as data. After all are entered, retrieve all; compare output to original. Close file, reopen, do retrieve and re-verify. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test013 Partial put test Overwrite entire records using partial puts. Make sure that NOOVERWRITE flag works. 1. Insert 10000 keys and retrieve them (equal key/data pairs). 2. Attempt to overwrite keys with NO_OVERWRITE set (expect error). 3. Actually overwrite each one with its datum reversed. No partial testing here. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test014 Exercise partial puts on short data Run 5 combinations of numbers of characters to replace, and number of times to increase the size by. Partial put test, small data, replacing with same size. The data set consists of the first nentries of the dictionary. We will insert them (and retrieve them) as we do in test 1 (equal key/data pairs). Then we'll try to perform partial puts of some characters at the beginning, some at the end, and some at the middle. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test015 Partial put test Partial put test where the key does not initially exist. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test016 Partial put test Partial put where the datum gets shorter as a result of the put. Partial put test where partial puts make the record smaller. Use the first 10,000 entries from the dictionary. Insert each with self as key and a fixed, medium length data string; retrieve each. After all are entered, go back and do partial puts, replacing a random-length string with the key value. Then verify. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test017 Basic offpage duplicate test. Run duplicates with small page size so that we test off page duplicates. Then after we have an off-page database, test with overflow pages too. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test018 Offpage duplicate test Key_{first,last,before,after} offpage duplicates. Run duplicates with small page size so that we test off page duplicates. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test019 Partial get test. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test020 In-Memory database tests. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test021 Btree range tests. Use the first 10,000 entries from the dictionary. Insert each with self, reversed as key and self as data. After all are entered, retrieve each using a cursor SET_RANGE, and getting about 20 keys sequentially after it (in some cases we'll run out towards the end of the file). =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test022 Test of DB->getbyteswapped(). =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test023 Duplicate test Exercise deletes and cursor operations within a duplicate set. Add a key with duplicates (first time on-page, second time off-page) Number the dups. Delete dups and make sure that CURRENT/NEXT/PREV work correctly. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test024 Record number retrieval test. Test the Btree and Record number get-by-number functionality. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test025 DB_APPEND flag test. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test026 Small keys/medium data w/duplicates Put/get per key. Loop through keys -- delete each key ... test that cursors delete duplicates correctly Keyed delete test through cursor. If ndups is small; this will test on-page dups; if it's large, it will test off-page dups. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test027 Off-page duplicate test Test026 with parameters to force off-page duplicates. Check that delete operations work. Create a database; close database and reopen it. Then issues delete by key for each entry. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test028 Cursor delete test Test put operations after deleting through a cursor. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test029 Test the Btree and Record number renumbering. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test030 Test DB_NEXT_DUP Functionality. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test031 Duplicate sorting functionality Make sure DB_NODUPDATA works. Use the first 10,000 entries from the dictionary. Insert each with self as key and "ndups" duplicates For the data field, prepend random five-char strings (see test032) that we force the duplicate sorting code to do something. Along the way, test that we cannot insert duplicate duplicates using DB_NODUPDATA. By setting ndups large, we can make this an off-page test After all are entered, retrieve all; verify output. Close file, reopen, do retrieve and re-verify. This does not work for recno =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test032 DB_GET_BOTH, DB_GET_BOTH_RANGE Use the first 10,000 entries from the dictionary. Insert each with self as key and "ndups" duplicates. For the data field, prepend the letters of the alphabet in a random order so we force the duplicate sorting code to do something. By setting ndups large, we can make this an off-page test. By setting overflow to be 1, we can make this an overflow test. Test the DB_GET_BOTH functionality by retrieving each dup in the file explicitly. Test the DB_GET_BOTH_RANGE functionality by retrieving the unique key prefix (cursor only). Finally test the failure case. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test033 DB_GET_BOTH without comparison function Use the first 10,000 entries from the dictionary. Insert each with self as key and data; add duplicate records for each. After all are entered, retrieve all and verify output using DB_GET_BOTH (on DB and DBC handles) and DB_GET_BOTH_RANGE (on a DBC handle) on existent and nonexistent keys. XXX This does not work for rbtree. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test034 test032 with off-page or overflow case with non-duplicates and duplicates. DB_GET_BOTH, DB_GET_BOTH_RANGE functionality with off-page or overflow case within non-duplicates and duplicates. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test035 Test033 with off-page non-duplicates and duplicates DB_GET_BOTH functionality with off-page non-duplicates and duplicates. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test036 Test KEYFIRST and KEYLAST when the key doesn't exist Put nentries key/data pairs (from the dictionary) using a cursor and KEYFIRST and KEYLAST (this tests the case where use use cursor put for non-existent keys). =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test037 Test DB_RMW =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test038 DB_GET_BOTH, DB_GET_BOTH_RANGE on deleted items Use the first 10,000 entries from the dictionary. Insert each with self as key and "ndups" duplicates. For the data field, prepend the letters of the alphabet in a random order so we force the duplicate sorting code to do something. By setting ndups large, we can make this an off-page test Test the DB_GET_BOTH and DB_GET_BOTH_RANGE functionality by retrieving each dup in the file explicitly. Then remove each duplicate and try the retrieval again. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test039 DB_GET_BOTH/DB_GET_BOTH_RANGE on deleted items without comparison function. Use the first 10,000 entries from the dictionary. Insert each with self as key and "ndups" duplicates. For the data field, prepend the letters of the alphabet in a random order so we force the duplicate sorting code to do something. By setting ndups large, we can make this an off-page test. Test the DB_GET_BOTH and DB_GET_BOTH_RANGE functionality by retrieving each dup in the file explicitly. Then remove each duplicate and try the retrieval again. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test040 Test038 with off-page duplicates DB_GET_BOTH functionality with off-page duplicates. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test041 Test039 with off-page duplicates DB_GET_BOTH functionality with off-page duplicates. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test042 Concurrent Data Store test (CDB) Multiprocess DB test; verify that locking is working for the concurrent access method product. Use the first "nentries" words from the dictionary. Insert each with self as key and a fixed, medium length data string. Then fire off multiple processes that bang on the database. Each one should try to read and write random keys. When they rewrite, they'll append their pid to the data string (sometimes doing a rewrite sometimes doing a partial put). Some will use cursors to traverse through a few keys before finding one to write. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test043 Recno renumbering and implicit creation test Test the Record number implicit creation and renumbering options. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test044 Small system integration tests Test proper functioning of the checkpoint daemon, recovery, transactions, etc. System integration DB test: verify that locking, recovery, checkpoint, and all the other utilities basically work. The test consists of $nprocs processes operating on $nfiles files. A transaction consists of adding the same key/data pair to some random number of these files. We generate a bimodal distribution in key size with 70% of the keys being small (1-10 characters) and the remaining 30% of the keys being large (uniform distribution about mean $key_avg). If we generate a key, we first check to make sure that the key is not already in the dataset. If it is, we do a lookup. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test045 Small random tester Runs a number of random add/delete/retrieve operations. Tests both successful conditions and error conditions. Run the random db tester on the specified access method. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test046 Overwrite test of small/big key/data with cursor checks. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test047 DBcursor->c_get get test with SET_RANGE option. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test048 Cursor stability across Btree splits. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test049 Cursor operations on uninitialized cursors. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test050 Overwrite test of small/big key/data with cursor checks for Recno. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test051 Fixed-length record Recno test. 0. Test various flags (legal and illegal) to open 1. Test partial puts where dlen != size (should fail) 2. Partial puts for existent record -- replaces at beg, mid, and end of record, as well as full replace =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test052 Renumbering record Recno test. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test053 Test of the DB_REVSPLITOFF flag in the Btree and Btree-w-recnum methods. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test054 Cursor maintenance during key/data deletion. This test checks for cursor maintenance in the presence of deletes. There are N different scenarios to tests: 1. No duplicates. Cursor A deletes a key, do a GET for the key. 2. No duplicates. Cursor is positioned right before key K, Delete K, do a next on the cursor. 3. No duplicates. Cursor is positioned on key K, do a regular delete of K, do a current get on K. 4. Repeat 3 but do a next instead of current. 5. Duplicates. Cursor A is on the first item of a duplicate set, A does a delete. Then we do a non-cursor get. 6. Duplicates. Cursor A is in a duplicate set and deletes the item. do a delete of the entire Key. Test cursor current. 7. Continue last test and try cursor next. 8. Duplicates. Cursor A is in a duplicate set and deletes the item. Cursor B is in the same duplicate set and deletes a different item. Verify that the cursor is in the right place. 9. Cursors A and B are in the place in the same duplicate set. A deletes its item. Do current on B. 10. Continue 8 and do a next on B. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test055 Basic cursor operations. This test checks basic cursor operations. There are N different scenarios to tests: 1. (no dups) Set cursor, retrieve current. 2. (no dups) Set cursor, retrieve next. 3. (no dups) Set cursor, retrieve prev. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test056 Cursor maintenance during deletes. Check if deleting a key when a cursor is on a duplicate of that key works. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test057 Cursor maintenance during key deletes. 1. Delete a key with a cursor. Add the key back with a regular put. Make sure the cursor can't get the new item. 2. Put two cursors on one item. Delete through one cursor, check that the other sees the change. 3. Same as 2, with the two cursors on a duplicate. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test058 Verify that deleting and reading duplicates results in correct ordering. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test059 Cursor ops work with a partial length of 0. Make sure that we handle retrieves of zero-length data items correctly. The following ops, should allow a partial data retrieve of 0-length. db_get db_cget FIRST, NEXT, LAST, PREV, CURRENT, SET, SET_RANGE =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test060 Test of the DB_EXCL flag to DB->open(). 1) Attempt to open and create a nonexistent database; verify success. 2) Attempt to reopen it; verify failure. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test061 Test of txn abort and commit for in-memory databases. a) Put + abort: verify absence of data b) Put + commit: verify presence of data c) Overwrite + abort: verify that data is unchanged d) Overwrite + commit: verify that data has changed e) Delete + abort: verify that data is still present f) Delete + commit: verify that data has been deleted =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test062 Test of partial puts (using DB_CURRENT) onto duplicate pages. Insert the first 200 words into the dictionary 200 times each with self as key and :self as data. Use partial puts to append self again to data; verify correctness. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test063 Test of the DB_RDONLY flag to DB->open Attempt to both DB->put and DBC->c_put into a database that has been opened DB_RDONLY, and check for failure. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test064 Test of DB->get_type Create a database of type specified by method. Make sure DB->get_type returns the right thing with both a normal and DB_UNKNOWN open. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test065 Test of DB->stat, both -DB_FAST_STAT and row counts with DB->stat -txn. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test066 Test of cursor overwrites of DB_CURRENT w/ duplicates. Make sure a cursor put to DB_CURRENT acts as an overwrite in a database with duplicates. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test067 Test of DB_CURRENT partial puts onto almost empty duplicate pages, with and without DB_DUP_SORT. Test of DB_CURRENT partial puts on almost-empty duplicate pages. This test was written to address the following issue, #2 in the list of issues relating to bug #0820: 2. DBcursor->put, DB_CURRENT flag, off-page duplicates, hash and btree: In Btree, the DB_CURRENT overwrite of off-page duplicate records first deletes the record and then puts the new one -- this could be a problem if the removal of the record causes a reverse split. Suggested solution is to acquire a cursor to lock down the current record, put a new record after that record, and then delete using the held cursor. It also tests the following, #5 in the same list of issues: 5. DBcursor->put, DB_AFTER/DB_BEFORE/DB_CURRENT flags, DB_DBT_PARTIAL set, duplicate comparison routine specified. The partial change does not change how data items sort, but the record to be put isn't built yet, and that record supplied is the one that's checked for ordering compatibility. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test068 Test of DB_BEFORE and DB_AFTER with partial puts. Make sure DB_BEFORE and DB_AFTER work properly with partial puts, and check that they return EINVAL if DB_DUPSORT is set or if DB_DUP is not. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test069 Test of DB_CURRENT partial puts without duplicates-- test067 w/ small ndups to ensure that partial puts to DB_CURRENT work correctly in the absence of duplicate pages. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test070 Test of DB_CONSUME (Four consumers, 1000 items.) Fork off six processes, four consumers and two producers. The producers will each put 20000 records into a queue; the consumers will each get 10000. Then, verify that no record was lost or retrieved twice. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test071 Test of DB_CONSUME (One consumer, 10000 items.) This is DB Test 70, with one consumer, one producers, and 10000 items. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test072 Test of cursor stability when duplicates are moved off-page. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test073 Test of cursor stability on duplicate pages. Does the following: a. Initialize things by DB->putting ndups dups and setting a reference cursor to point to each. b. c_put ndups dups (and correspondingly expanding the set of reference cursors) after the last one, making sure after each step that all the reference cursors still point to the right item. c. Ditto, but before the first one. d. Ditto, but after each one in sequence first to last. e. Ditto, but after each one in sequence from last to first. occur relative to the new datum) f. Ditto for the two sequence tests, only doing a DBC->c_put(DB_CURRENT) of a larger datum instead of adding a new one. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test074 Test of DB_NEXT_NODUP. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test076 Test creation of many small databases in a single environment. [#1528]. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test077 Test of DB_GET_RECNO [#1206]. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test078 Test of DBC->c_count(). [#303] =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test079 Test of deletes in large trees. (test006 w/ sm. pagesize). Check that delete operations work in large btrees. 10000 entries and a pagesize of 512 push this out to a four-level btree, with a small fraction of the entries going on overflow pages. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test081 Test off-page duplicates and overflow pages together with very large keys (key/data as file contents). =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test082 Test of DB_PREV_NODUP (uses test074). =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test083 Test of DB->key_range. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test084 Basic sanity test (test001) with large (64K) pages. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test085 Test of cursor behavior when a cursor is pointing to a deleted btree key which then has duplicates added. [#2473] =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test086 Test of cursor stability across btree splits/rsplits with subtransaction aborts (a variant of test048). [#2373] =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test087 Test of cursor stability when converting to and modifying off-page duplicate pages with subtransaction aborts. [#2373] Does the following: a. Initialize things by DB->putting ndups dups and setting a reference cursor to point to each. Do each put twice, first aborting, then committing, so we're sure to abort the move to off-page dups at some point. b. c_put ndups dups (and correspondingly expanding the set of reference cursors) after the last one, making sure after each step that all the reference cursors still point to the right item. c. Ditto, but before the first one. d. Ditto, but after each one in sequence first to last. e. Ditto, but after each one in sequence from last to first. occur relative to the new datum) f. Ditto for the two sequence tests, only doing a DBC->c_put(DB_CURRENT) of a larger datum instead of adding a new one. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test088 Test of cursor stability across btree splits with very deep trees (a variant of test048). [#2514] =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test089 Concurrent Data Store test (CDB) Enhanced CDB testing to test off-page dups, cursor dups and cursor operations like c_del then c_get. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test090 Test for functionality near the end of the queue using test001. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test091 Test of DB_CONSUME_WAIT. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test092 Test of DB_DIRTY_READ [#3395] We set up a database with nentries in it. We then open the database read-only twice. One with dirty reads and one without. We open the database for writing and update some entries in it. Then read those new entries via db->get (clean and dirty), and via cursors (clean and dirty). =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test093 Test set_bt_compare (btree key comparison function) and set_h_compare (hash key comparison function). Open a database with a comparison function specified, populate, and close, saving a list with that key order as we do so. Reopen and read in the keys, saving in another list; the keys should be in the order specified by the comparison function. Sort the original saved list of keys using the comparison function, and verify that it matches the keys as read out of the database. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test094 Test using set_dup_compare. Use the first 10,000 entries from the dictionary. Insert each with self as key and data; retrieve each. After all are entered, retrieve all; compare output to original. Close file, reopen, do retrieve and re-verify. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test095 Bulk get test for methods supporting dups. [#2934] =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test096 Db->truncate test. For all methods: Test that truncate empties an existing database. Test that truncate-write in an aborted txn doesn't change the original contents. Test that truncate-write in a committed txn does overwrite the original contents. For btree and hash, do the same in a database with offpage dups. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test097 Open up a large set of database files simultaneously. Adjust for local file descriptor resource limits. Then use the first 1000 entries from the dictionary. Insert each with self as key and a fixed, medium length data string; retrieve each. After all are entered, retrieve all; compare output to original. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test098 Test of DB_GET_RECNO and secondary indices. Open a primary and a secondary, and do a normal cursor get followed by a get_recno. (This is a smoke test for "Bug #1" in [#5811].) =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test099 Test of DB->get and DBC->c_get with set_recno and get_recno. Populate a small btree -recnum database. After all are entered, retrieve each using -recno with DB->get. Open a cursor and do the same for DBC->c_get with set_recno. Verify that set_recno sets the record number position properly. Verify that get_recno returns the correct record numbers. Using the same database, open 3 cursors and position one at the beginning, one in the middle, and one at the end. Delete by cursor and check that record renumbering is done properly. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test100 Test for functionality near the end of the queue using test025 (DB_APPEND). =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test101 Test for functionality near the end of the queue using test070 (DB_CONSUME). =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test102 Bulk get test for record-based methods. [#2934] =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test103 Test bulk get when record numbers wrap around. Load database with items starting before and ending after the record number wrap around point. Run bulk gets (-multi_key) with various buffer sizes and verify the contents returned match the results from a regular cursor get. Then delete items to create a sparse database and make sure it still works. Test both -multi and -multi_key since they behave differently. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test106 =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test107 Test of read-committed (degree 2 isolation). [#8689] We set up a database. Open a read-committed transactional cursor and a regular transactional cursor on it. Position each cursor on one page, and do a put to a different page. Make sure that: - the put succeeds if we are using degree 2 isolation. - the put deadlocks within a regular transaction with a regular cursor. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test109 Test of sequences. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test110 Partial get test with duplicates. For hash and btree, create and populate a database with dups. Randomly selecting offset and length, retrieve data from each record and make sure we get what we expect. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test111 Test database compaction. Populate a database. Remove a high proportion of entries. Dump and save contents. Compact the database, dump again, and make sure we still have the same contents. Add back some entries, delete more entries (this time by cursor), dump, compact, and do the before/after check again. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test112 Test database compaction with a deep tree. This is a lot like test111, but with a large number of entries and a small page size to make the tree deep. To make it simple we use numerical keys all the time. Dump and save contents. Compact the database, dump again, and make sure we still have the same contents. Add back some entries, delete more entries (this time by cursor), dump, compact, and do the before/after check again. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test113 Test database compaction with duplicates. This is essentially test111 with duplicates. To make it simple we use numerical keys all the time. Dump and save contents. Compact the database, dump again, and make sure we still have the same contents. Add back some entries, delete more entries (this time by cursor), dump, compact, and do the before/after check again. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test114 Test database compaction with overflows. Populate a database. Remove a high proportion of entries. Dump and save contents. Compact the database, dump again, and make sure we still have the same contents. Add back some entries, delete more entries (this time by cursor), dump, compact, and do the before/after check again. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test115 Test database compaction with user-specified btree sort. This is essentially test111 with the user-specified sort. Populate a database. Remove a high proportion of entries. Dump and save contents. Compact the database, dump again, and make sure we still have the same contents. Add back some entries, delete more entries (this time by cursor), dump, compact, and do the before/after check again. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test116 Test of basic functionality of lsn_reset. Create a database in an env. Copy it to a new file within the same env. Reset the page LSNs. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test117 Test database compaction with requested fill percent. Populate a database. Remove a high proportion of entries. Dump and save contents. Compact the database, requesting fill percentages starting at 10% and working our way up to 100. On each cycle, make sure we still have the same contents. Unlike the other compaction tests, this one does not use -freespace. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test119 Test behavior when Berkeley DB returns DB_BUFFER_SMALL on a cursor. If the user-supplied buffer is not large enough to contain the returned value, DB returns BUFFER_SMALL. If it does, check that the cursor does not move -- if it moves, it will skip items. [#13815] =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test120 Test of multi-version concurrency control. Test basic functionality: a snapshot transaction started before a regular transaction's put can't see the modification. A snapshot transaction started after the put can see it. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test121 Tests of multi-version concurrency control. MVCC and cursor adjustment. Set up a -snapshot cursor and position it in the middle of a database. Write to the database, both before and after the cursor, and verify that it stays on the same position. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test122 Tests of multi-version concurrency control. MVCC and databases that turn multi-version on and off. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test123 Concurrent Data Store cdsgroup smoke test. Open a CDS env with -cdb_alldb. Start a "txn" with -cdsgroup. Create two databases in the env, do a cursor put in both within the same txn. This should succeed. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test124 Test db->verify with noorderchk and orderchkonly flags. Create a db with a non-standard sort order. Check that it fails a regular verify and succeeds with -noorderchk. Do a similar test with a db containing subdbs, one with the standard order and another with non-standard. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test125 Test cursor comparison API. The cursor comparison API reports whether two cursors within the same database are at the same position. It does not report any information about relative position. 1. Test two uninitialized cursors (error). 2. Test one uninitialized cursor, one initialized (error). 3. Test two cursors in different databases (error). 4. Put two cursors in the same place, test for match. Walk them back and forth a bit, more matching. 5. Two cursors in the same spot. Delete through one. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test126 Test database bulk update for non-duplicate databases. Put with -multiple, then with -multiple_key, and make sure the items in database are what we put. Later, delete some items with -multiple, then with -multiple_key, and make sure if the correct items are deleted. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test127 Test database bulk update. This is essentially test126 with duplicates. To make it simple we use numerical keys all the time. Put with -multiple, then with -multiple_key, and make sure the items in database are what we want. Later, delete some items with -multiple, then with -multiple_key, and make sure if the correct items are deleted. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test128 Test database bulk update for non-duplicate databases, with different configurations. This is essentially test126 with the following configurations: * sub database. * secondary database. * bulk buffer pre-sort. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test129 Test database bulk update for duplicate database, with different configurations. This is essentially test127 with the following configurations: * sub database. * bulk buffer pre-sort. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test130 Test moving of subdatabase metadata pages. Populate num_db sub-database. Open multiple handles on each. Remove a high proportion of entries. Dump and save contents. Compact the database, dump again, and make sure we still have the same contents. Make sure handles and cursors still work after compaction. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test131 Test foreign database operations. Create a foreign db, and put some records into it. Then associate the foreign db with a secondary db, and put records into the primary db. Do operations in the foreign db and check results. Finally, verify the foreign relation between the foreign db and secondary db. Here, we test three different foreign delete constraints: - DB_FOREIGN_ABORT - DB_FOREIGN_CASCADE - DB_FOREIGN_NULLIFY =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test132 Test foreign database operations on sub databases and in-memory databases. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test133 Test Cursor Cleanup. Open a primary database and a secondary database, then open 3 cursors on the secondary database, and point them at the first item. Do the following operations in loops: * The 1st cursor will delete the current item. * The 2nd cursor will also try to delete the current item. * Move all the 3 cursors to get the next item and check the returns. Finally, move the 3rd cursor once. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test134 Test cursor cleanup for sub databases. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test135 Test operations on similar overflow records. [#20329] Open 3 kinds of databases: one is not dup, and one is dup-unsorted while another one is dup-sorted. Then we put some similar but not identical overflow records into these databases. Also, we try to remove some records from these databases. We'll verify the records after put and after deletion. Here is how we define the 'similar overflow' in this test: * Both the key.size and data.size are not small than DB's pagesize: The key.size is around 2*pagesize and data.size equals to pagesize. * The keys are of same length, and they only differ in a small piece. The location of difference could be in the start/middle/end. * For dup databases, the dup datas have same rule as the key rule mentioned above. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test136 Test operations on similar overflow records. [#20329] Here, we use subdatabases. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test137 Test Automatic Resource Management. [#16188][#20281] Open an environment, and open a database in it. Do some operations in the database, including: insert, dump, delete Close the environment without closing the database. Re-open the database and verify the records are the ones we expected. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test138 Test Automatic Resource Management. [#16188][#20281] Here, we test the following cases: Non-encrypt for cds Non-encrypt for tds Encrypt for ds Encrypt for cds Encrypt for tds =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test139 Verify an open database. Create and populate a database, leave open, and run db->verify. Delete half the data, verify, compact, verify again. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test140 Test expansion and contraction of hash chains with deletion, compaction, and recovery. On each cycle, populate a database to create reasonably long hash chains, say 8 - 10 entries. Delete some entries, check for consistency, compact, and check for consistency again. We both commit and abort the compact, and in the case of the abort, compare the recovered database to the pre-compact database. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test141 Test moving a large part of a hash table to a continguous block of free pages. Create a database with a btree and a hash subdb. Remove all the data from the first (btree) db and then compact to cause the hash db to get moved. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= test142 Tests exclusive database handles. 1. Test that exclusive database handles return an error in incompatible environments. 2. Test that an exclusive subdatabase handle does not block opening a handle on another database in the same file. 3. Run script tests on a subdatabase with both wait and no wait configuration. 4. Run script tests on a database with both wait and no wait configuration. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= txn001 Begin, commit, abort testing. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= txn002 Verify that read-only transactions do not write log records. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= txn003 Test abort/commit/prepare of txns with outstanding child txns. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= txn004 Test of wraparound txnids (txn001) =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= txn005 Test transaction ID wraparound and recovery. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= txn008 Test of wraparound txnids (txn002) =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= txn009 Test of wraparound txnids (txn003) =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= txn010 Test DB_ENV->txn_checkpoint arguments/flags =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= txn011 Test durable and non-durable txns. Test a mixed env (with both durable and non-durable dbs), then a purely non-durable env. Make sure commit and abort work, and that only the log records we expect are written. Test that we can't get a durable handle on an open ND database, or vice versa. Test that all subdb's must be of the same type (D or ND). =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= txn012 Test txn->getname and txn->setname. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= txn013 Test of txns used in the wrong environment. Set up two envs. Start a txn in one env, and attempt to use it in the other env. Verify we get the appropriate error message. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= txn014 Test of parent and child txns working on the same database. A txn that will become a parent create a database. A txn that will not become a parent creates another database. Start a child txn of the 1st txn. Verify that the parent txn is disabled while child is open. 1. Child reads contents with child handle (should succeed). 2. Child reads contents with parent handle (should succeed). Verify that the non-parent txn can read from its database, and that the child txn cannot. Return to the child txn. 3. Child writes with child handle (should succeed). 4. Child writes with parent handle (should succeed). Commit the child, verify that the parent can write again. Check contents of database with a second child.