[ZODB-Dev] PopulatingLargeDB
Tamas Hegedus
hegedus at med.unc.edu
Fri Feb 10 15:43:56 EST 2006
Hi,
I am using ZODB3-3.5.1. I would like to populate it with large number of
objects (n>100,000). I try to use subtransaction/savepoint to save
memory after every 10,000th object (my objects have nested objects)
I see the increase of the size of the Data.fs file after commits, but my
memory does not seem to be freed: the red line just goes upper and upper
while my disk cash (swap) also started to be used heavily... I have 1G
RAM...
The size of the file I parse into objects is almost 798M, approx with
200,000 records.
Parts of the code I use listed below (I tried different combinations of
commit/savepoint/etc; but I do not really understand the whole picture...).
Thanks for your help and for any suggestion,
Tamas
#==============================================================
db = ZODB.config.databaseFromURL("etc/zodb.conf")
connection = db.open()
droot = connection.root()
droot['MyObjs'] = OOBTree()
myDb = droot['MyObjs']
#--------------------------------------------------------------
MixIn( MySP.Record, Persistent)
it = MySP.Iterator( open( 'dat_file.txt'),
MySP.RecordParser())
i = 0
for rec in it:
id = copy.deepcopy( rec.ids[0])
myDb[id] = rec
i += 1
if i % 10000 == 0:
transaction.savepoint(True)
#transaction.commit()
#transaction.manager.free( transaction.get())
print i
transaction.commit()
connection.close()
#==============================================================
--
Tamas Hegedus, PhD | phone: (1) 919-966 0329
UNC - Biochem & Biophys | fax: (1) 919-966 5178
5007A Thurston-Bowles Bldg | mailto:hegedus at med.unc.edu
Chapel Hill, NC, 27599-7248 | http://biohegedus.org
More information about the ZODB-Dev
mailing list