[Zodb-checkins] CVS: Zope3/src/zodb/storage - memory.py:1.2
interfaces.py:1.17 fsdump.py:1.6 bdbminimal.py:1.18
bdbfull.py:1.24 base.py:1.27 fsindex.py:NONE file.py:NONE
Jeremy Hylton
jeremy at zope.com
Tue Apr 22 12:23:15 EDT 2003
Update of /cvs-repository/Zope3/src/zodb/storage
In directory cvs.zope.org:/tmp/cvs-serv13741
Modified Files:
memory.py interfaces.py fsdump.py bdbminimal.py bdbfull.py
base.py
Removed Files:
fsindex.py file.py
Log Message:
Merge the jeremy-new-pack-branch to the trunk.
The primary change is a completely new implementation of file storage pack.
=== Zope3/src/zodb/storage/memory.py 1.1 => 1.2 ===
--- Zope3/src/zodb/storage/memory.py:1.1 Thu Mar 20 17:58:16 2003
+++ Zope3/src/zodb/storage/memory.py Tue Apr 22 11:23:13 2003
@@ -224,8 +224,8 @@
raise db.DBNotFoundError
while self._keys[i] == key and self._vals[i] <> val:
i += 1
- if i > len(self):
- raise db.DBNotFoundError
+ if i >= len(self):
+ raise db.DBNotFoundError
return self._getrec(i)
def set(self, key):
@@ -293,3 +293,6 @@
def append(self, val, txn=None):
super(FakeQueue, self).append(val)
+
+ def values(self):
+ return self
=== Zope3/src/zodb/storage/interfaces.py 1.16 => 1.17 ===
--- Zope3/src/zodb/storage/interfaces.py:1.16 Tue Apr 8 10:45:55 2003
+++ Zope3/src/zodb/storage/interfaces.py Tue Apr 22 11:23:13 2003
@@ -186,8 +186,8 @@
incremental pack, only old object revisions are removed. In a full gc
pack, cyclic garbage detection and removal is also performed.
- t is the pack time. All non-current object revisions older than t
- will be removed in an incremental pack.
+ t is the pack time. All non-current object revisions older than
+ or the same age as t will be removed in an incremental pack.
pack() always performs an incremental pack. If the gc flag is True,
then pack() will also perform a garbage collection. Some storages
=== Zope3/src/zodb/storage/fsdump.py 1.5 => 1.6 ===
--- Zope3/src/zodb/storage/fsdump.py:1.5 Mon Mar 17 15:18:27 2003
+++ Zope3/src/zodb/storage/fsdump.py Tue Apr 22 11:23:13 2003
@@ -11,98 +11,4 @@
# FOR A PARTICULAR PURPOSE.
#
##############################################################################
-"""A low-level utility to dump the internal FileStorage representation."""
-
-import struct
-from zodb.storage.file \
- import TRANS_HDR, TRANS_HDR_LEN, DATA_HDR, DATA_HDR_LEN
-from zodb.utils import u64
-from zodb.storage.base import splitrefs
-from zodb.storage.tests.base import zodb_unpickle
-
-def fmt(p64):
- # Return a nicely formatted string for a packaged 64-bit value
- return "%016x" % u64(p64)
-
-def dump(path, dest=None):
- Dumper(path, dest).dump()
-
-class Dumper:
- """A very verbose dumper for debugging FileStorage problems."""
-
- def __init__(self, path, dest=None):
- self.file = open(path, "rb")
- self.dest = dest
-
- def dump(self):
- fid = self.file.read(1024)
- print >> self.dest, "*" * 60
- print >> self.dest, "file identifier: %r" % fid[:4]
- print >> self.dest, "database version: %r" % fid[4:8]
- # XXX perhaps verify that the rest of the metadata is nulls?
- while self.dump_txn():
- pass
-
- def dump_txn(self):
- pos = self.file.tell()
- h = self.file.read(TRANS_HDR_LEN)
- if not h:
- return False
- tid, tlen, status, ul, dl, el = struct.unpack(TRANS_HDR, h)
- end = pos + tlen
- print >> self.dest, "=" * 60
- print >> self.dest, "offset: %d" % pos
- print >> self.dest, "end pos: %d" % end
- print >> self.dest, "transaction id: %s" % fmt(tid)
- print >> self.dest, "trec len: %d" % tlen
- print >> self.dest, "status: %r" % status
- user = descr = extra = ""
- if ul:
- user = self.file.read(ul)
- if dl:
- descr = self.file.read(dl)
- if el:
- extra = self.file.read(el)
- print >> self.dest, "user: %r" % user
- print >> self.dest, "description: %r" % descr
- print >> self.dest, "len(extra): %d" % el
- while self.file.tell() < end:
- self.dump_data(pos)
- tlen2 = u64(self.file.read(8))
- print >> self.dest, "redundant trec len: %d" % tlen2
- return True
-
- def dump_data(self, tloc):
- pos = self.file.tell()
- h = self.file.read(DATA_HDR_LEN)
- assert len(h) == DATA_HDR_LEN
- oid, revid, prev, tloc, vlen, nrefs, dlen = struct.unpack(DATA_HDR, h)
- print >> self.dest, "-" * 60
- print >> self.dest, "offset: %d" % pos
- print >> self.dest, "oid: %s" % fmt(oid)
- print >> self.dest, "revid: %s" % fmt(revid)
- print >> self.dest, "previous record offset: %d" % prev
- print >> self.dest, "transaction offset: %d" % tloc
- if vlen:
- pnv = self.file.read(8)
- sprevdata = self.file.read(8)
- version = self.file.read(vlen)
- print >> self.dest, "version: %r" % version
- print >> self.dest, "non-version data offset: %d" % u64(pnv)
- print >> self.dest, \
- "previous version data offset: %d" % u64(sprevdata)
- print >> self.dest, 'numrefs:', nrefs
- for ref in splitrefs(self.file.read(nrefs * 8)):
- print >> self.dest, '\t%s' % fmt(ref)
- print >> self.dest, "len(data): %d" % dlen
- data = self.file.read(dlen)
- # A debugging feature for use with the test suite.
- if data.startswith("(czodb.storage.tests.minpo\nMinPO\n"):
- print >> self.dest, "value: %r" % zodb_unpickle(data).value
- if not dlen:
- sbp = self.file.read(8)
- print >> self.dest, "backpointer: %d" % u64(sbp)
-
-if __name__ == "__main__":
- import sys
- Dumper(sys.argv[1]).dump()
+from zodb.storage.file.dump import *
=== Zope3/src/zodb/storage/bdbminimal.py 1.17 => 1.18 ===
--- Zope3/src/zodb/storage/bdbminimal.py:1.17 Wed Apr 9 13:58:20 2003
+++ Zope3/src/zodb/storage/bdbminimal.py Tue Apr 22 11:23:13 2003
@@ -23,6 +23,8 @@
from zodb.conflict import ResolvedSerial
from zodb.storage.base import db, BerkeleyBase, PackStop, _WorkThread
from zodb.storage.base import splitrefs
+# For debugging
+from zodb.interfaces import _fmt_oid as fo
ABORT = 'A'
COMMIT = 'C'
@@ -101,7 +103,7 @@
# It is also used during pack to list objects for which no more
# references exist, such that the objects can be completely packed
# away.
- pass
+ self._oidqueue = self._setupDB('oidqueue', 0, db.DB_QUEUE, 8)
def _version_check(self, txn):
version = self._info.get('version')
@@ -472,6 +474,7 @@
finally:
c.close()
# We're done with the mark table
+ self._packmark.truncate(txn=txn)
def _collect_objs(self, txn):
orec = self._oidqueue.consume(txn)
=== Zope3/src/zodb/storage/bdbfull.py 1.23 => 1.24 ===
--- Zope3/src/zodb/storage/bdbfull.py:1.23 Thu Apr 10 15:06:53 2003
+++ Zope3/src/zodb/storage/bdbfull.py Tue Apr 22 11:23:13 2003
@@ -31,6 +31,8 @@
from zodb.storage.base import db, BerkeleyBase, PackStop, _WorkThread, \
splitrefs
from zodb.storage._helper import incr
+# For debugging
+from zodb.interfaces import _fmt_oid as fo
ABORT = 'A'
COMMIT = 'C'
@@ -120,7 +122,7 @@
# pending table is empty, the oids, pvids, and prevrevids tables
# must also be empty.
#
- # packmark -- [oid]
+ # packmark -- oid -> [tid]
# Every object reachable from the root during a classic pack
# operation will have its oid present in this table.
#
@@ -232,6 +234,7 @@
# Tables to support packing.
self._objrevs = self._setupDB('objrevs', db.DB_DUP)
self._delqueue = self._setupDB('delqueue', 0, db.DB_QUEUE, 8)
+ self._oidqueue = self._setupDB('oidqueue', 0, db.DB_QUEUE, 16)
def _version_check(self, txn):
version = self._info.get('version')
@@ -452,7 +455,7 @@
# created in the interrim.
if self._packing:
for oid in self._oids.keys():
- self._packmark.put(oid, PRESENT, txn=txn)
+ self._packmark.put(oid, tid, txn=txn)
self._oids.truncate(txn)
def _dobegin(self, txn, tid):
@@ -1422,8 +1425,6 @@
if self._metadata.has_key(orevid):
metadata = self._metadata[orevid]
self._metadata.delete(orevid, txn=txn)
- if self._references.has_key(orevid):
- self._references.delete(orevid, txn=txn)
# Decref the pickle
self._decrefPickle(oid, metadata[16:24], txn)
try:
@@ -1452,7 +1453,7 @@
refcount = u64(self._pickleRefcounts.get(revid, ZERO)) - 1
assert refcount >= 0
if refcount == 0:
- # We can collect this pickle
+ # We can collect this pickle and the references
self._pickleRefcounts.delete(revid, txn=txn)
self._pickles.delete(revid, txn=txn)
# And decref all objects pointed to by this pickle
@@ -1461,6 +1462,7 @@
deltas = {}
self._update(deltas, references, -1)
self._decref(deltas, txn)
+ self._references.delete(revid, txn=txn)
else:
self._pickleRefcounts.put(revid, p64(refcount), txn=txn)
@@ -1550,7 +1552,7 @@
# BAW: Maybe this could probably be more efficient by not doing so
# much searching, but it would also be more complicated, so the
# tradeoff should be measured.
- serial = None
+ serial, tid = self._getSerialAndTid(oid)
c = self._metadata.cursor(txn=txn)
try:
rec = c.set_range(oid)
@@ -1568,9 +1570,60 @@
c.close()
return serial
+ def _rootset(self, packtid, txn):
+ c = self._txnoids.cursor(txn)
+ try:
+ rec = c.first()
+ while rec:
+ tid, oid = rec
+ rec = c.next()
+ finally:
+ c.close()
+ # Find the root set for reachability purposes. A root set is a tuple
+ # of oid and tid. First, the current root object as of the pack time
+ # is always in the root set. Second, any object revision after the
+ # pack time that has a back pointer (lrevid) to before the pack time
+ # serves as another root because some future undo could then revive
+ # any referenced objects.
+ try:
+ zerorev = self._findrev(ZERO, packtid, txn)
+ except KeyError:
+ # There's no root object
+ return
+ self._oidqueue.append(ZERO+zerorev, txn)
+ c = self._txnoids.cursor(txn)
+ try:
+ try:
+ rec = c.set_range(packtid)
+ except db.DBNotFoundError:
+ rec = None
+ while rec:
+ tid, oid = rec
+ revid = oid + tid
+ rec = c.next()
+ lrevid = self._metadata[revid][16:24]
+ if lrevid < packtid:
+ self._oidqueue.append(revid, txn)
+ finally:
+ c.close()
+
+ # tid is None if all we care about is that any object revision is present.
+ def _packmark_has(self, oid, tid, txn):
+ if tid is None:
+ return self._packmark.has_key(oid)
+ c = self._packmark.cursor(txn)
+ try:
+ try:
+ c.set_both(oid, tid)
+ return True
+ except db.DBNotFoundError:
+ return False
+ finally:
+ c.close()
+
def _mark(self, txn, packtid):
# Find the oids for all the objects reachable from the root, as of the
- # pack time. To reduce the amount of in-core memory we need do do a
+ # pack time. To reduce the amount of in-core memory we need to do a
# pack operation, we'll save the mark data in the packmark table. The
# oidqueue is a BerkeleyDB Queue that holds the list of object ids to
# look at next, and by using this we don't need to keep an in-memory
@@ -1579,20 +1632,23 @@
# Quick exit for empty storages
if not self._serials:
return
- # The oid of the object we're looking at, starting at the root
- oid = ZERO
- # Start at the root, find all the objects the current revision of the
- # root references, and then for each of those, find all the objects it
- # references, and so on until we've traversed the entire object graph.
- while oid:
+ self._rootset(packtid, txn)
+ rec = self._oidqueue.consume(txn)
+ while rec:
if self._stop:
raise PackStop, 'stopped in _mark()'
- if not self._packmark.has_key(oid):
- # We haven't seen this object yet
- self._packmark.put(oid, PRESENT, txn=txn)
- # Get the list of references for the most current revision of
- # this object as of the pack time.
- tid = self._findrev(oid, packtid, txn)
+ revid = rec[1]
+ oid = revid[:8]
+ tid = revid[8:]
+ # See if this revision is already in the packmark
+ if not self._packmark_has(oid, tid, txn):
+ # BAW: We are more conservative than FileStorage here, since
+ # any reference to an object keeps all the object references
+ # alive. FileStorage will collect individual object
+ # revisions. I think our way is fine since we'll eventually
+ # collect everything incrementally anyway, and for Berkeley,
+ # all object revisions add to the refcount total.
+ self._packmark.put(oid, tid, txn=txn)
# Say there's no root object (as is the case in some of the
# unit tests), and we're looking up oid ZERO. Then serial
# will be None.
@@ -1602,11 +1658,13 @@
# object revision
references = self._references.get(oid+lrevid)
if references:
- for oid in splitrefs(references):
- self._oidqueue.append(oid, txn)
+ for roid in splitrefs(references):
+ # Find the most recent object revision as of the
+ # timestamp of the under-focus revision.
+ rtid = self._findrev(roid, tid, txn)
+ self._oidqueue.append(roid+rtid, txn)
# Pop the next oid off the queue and do it all again
rec = self._oidqueue.consume(txn)
- oid = rec and rec[1]
assert len(self._oidqueue) == 0
def _sweep(self, txn, packtid):
@@ -1627,7 +1685,7 @@
# Otherwise, if packmark (which knows about all the root
# reachable objects) doesn't have a record for this guy, then
# we can zap it. Do so by appending to oidqueue.
- if not self._packmark.has_key(oid):
+ if not self._packmark_has(oid, None, txn):
self._delqueue.append(oid, txn)
finally:
c.close()
=== Zope3/src/zodb/storage/base.py 1.26 => 1.27 ===
--- Zope3/src/zodb/storage/base.py:1.26 Wed Apr 9 13:54:51 2003
+++ Zope3/src/zodb/storage/base.py Tue Apr 22 11:23:13 2003
@@ -80,6 +80,7 @@
_vote()
_abort()
_finish()
+ _clear_temp()
If the subclass wants to implement IUndoStorage, it must implement
all the methods in that interface.
@@ -535,8 +536,7 @@
self._references = self._setupDB('references')
self._oids = self._setupDB('oids')
self._pending = self._setupDB('pending')
- self._packmark = self._setupDB('packmark')
- self._oidqueue = self._setupDB('oidqueue', 0, db.DB_QUEUE, 8)
+ self._packmark = self._setupDB('packmark', db.DB_DUP)
# Do storage specific initialization
self._init()
self._withtxn(self._version_check)
=== Removed File Zope3/src/zodb/storage/fsindex.py ===
=== Removed File Zope3/src/zodb/storage/file.py ===
More information about the Zodb-checkins
mailing list