Did I find the right examples for you? yes no

All Samples(41)  |  Call(22)  |  Derive(0)  |  Import(19)

src/o/p/openlibrary-HEAD/openlibrary/data/solr.py   openlibrary(Download)
def process_author_dump(writer, authors_dump):
    import bsddb 
    db = bsddb.btopen('solrdump/authors.db', 'w', cachesize=1024*1024*1024)
 
    properties = ['key', 'name', 'alternate_names', 'personal_name']
def process_redirect_dump(writer, redirects_dump):
    import bsddb 
    db = bsddb.btopen('solrdump/redirects.db', 'w', cachesize=1024*1024*1024)
 
    for type, key, revision, timestamp, json in read_tsv(redirects_dump):

src/o/p/openlibrary-HEAD/openlibrary/data/dump.py   openlibrary(Download)
def make_bsddb(dbfile, dump_file):
    import bsddb 
    db = bsddb.btopen(dbfile, 'w', cachesize=1024*1024*1024)
 
    from infogami.infobase.utils import flatten_dict

src/o/p/openlibrary-HEAD/scripts/2010/03/olload.py   openlibrary(Download)
def load_bsddb(filename, data):
    import bsddb
    GB = 1234 ** 3
    db = bsddb.btopen(filename, cachesize=1 * GB)
 

src/m/e/MessageQueue-HEAD/processors/diskqueueprocessor.py   MessageQueue(Download)
		self.file_hash = path + file_md5 + ext_hash
 
		self.db_tree = bsddb.btopen(  self.file_tree, "c")
		self.db_hash = bsddb.hashopen(self.file_hash, "c")
 
		try:
			self.db_tree.close()
			self.db_tree = bsddb.btopen(  self.file_tree, "n")
 
			self.db_hash.close()

src/b/l/blastkit-HEAD/lib/blastparser.py   blastkit(Download)
    def __init__(self, filename, mode='r'):
        from shelve import BsdDbShelf
        from bsddb import btopen
 
        _db = btopen(filename, 'r')
def open_shelf(filename, mode='r'):
    from shelve import BsdDbShelf
    from bsddb import btopen
 
    return _BlastShelf(filename, mode)
    import sys
    from shelve import BsdDbShelf
    from bsddb import btopen
    from optparse import OptionParser
 
        blast_fp = open(blast_file)
 
    _db = btopen(output_file, 'c')
    db = BsdDbShelf(_db)
 

src/a/p/Apache2Piwik-HEAD/apache2piwik.py   Apache2Piwik(Download)
SV_ON_DISK = False
if SV_ON_DISK:
    STILL_VISITING = bsddb.btopen(db, 'c')
else:
    if not 'still_visiting.db' in os.listdir('./data/cache'+str(ID_SITE)):
        STILL_VISITING = {}
    else:
        STILL_VISITING = {}
        SV = bsddb.btopen(db, 'c')
def store_still_visiting():
    sv = bsddb.btopen(db, 'c')
    for k in STILL_VISITING.keys():
        sv[k]=STILL_VISITING[k]
 
            global STILL_VISITING
            if SV_ON_DISK:
                STILL_VISITING = bsddb.btopen(db, 'c')
            else:
                STILL_VISITING = {}

src/d/u/dupfilefind-1.6.9/dupfilefind/dff.py   dupfilefind(Download)
    tempd = fileutil.NamedTemporaryDirectory()
    db1fname = os.path.join(tempd.name, TMPDB1FNAME)
    d1 = bsddb.btopen(db1fname, "n")
    tempd.register_file(d1)
    fileutil.remove_if_possible(db1fname)
    db2fname = os.path.join(tempd.name, TMPDB2FNAME)
    d2 = bsddb.btopen(db2fname, "n")

src/j/c/jcvi-HEAD/formats/fileindex.py   jcvi(Download)
        if need_update(filename, self.idxfile):
            self.clear()
            self.db = bsddb.btopen(self.idxfile, 'c')
            self.create()
            self.db.close()
 
        self.db = bsddb.btopen(self.idxfile, 'r')

src/m/e/MessageQueue-HEAD/processors/singlediskqueueprocessor.py   MessageQueue(Download)
		if db.get(self.dbname) is None:
			# Open the database
			db[self.dbname] = bsddb.btopen(self.dbname, "c")
 
		# Collect a copy of the db

src/m/e/MessageQueue-HEAD/processors/randomdiskqueueprocessor.py   MessageQueue(Download)
		if db.get(self.dbname) is None:
			# Open the database
			db[self.dbname] = bsddb.btopen(self.dbname, "c")
 
		# Collect a copy of the db

  1 | 2 | 3  Next