methods = []
urldata_cache = {}
saved_headrevs = {}
+persistent_database_connection = {}
def fetcher_init(d):
"""
Called to initialize the fetchers once the configuration data is known.
Calls before this must not hit the cache.
"""
- pd = persist_data.PersistData(d)
+ pd = persist_data.PersistData(d, persistent_database_connection)
# When to drop SCM head revisions controlled by user policy
srcrev_policy = bb.data.getVar('BB_SRCREV_POLICY', d, 1) or "clear"
if srcrev_policy == "cache":
return true/false on whether they've changed.
"""
- pd = persist_data.PersistData(d)
+ pd = persist_data.PersistData(d, persistent_database_connection)
data = pd.getKeyValues("BB_URI_HEADREVS")
data2 = bb.fetch.saved_headrevs
if not hasattr(self, "_latest_revision"):
raise ParameterError
- pd = persist_data.PersistData(d)
+ pd = persist_data.PersistData(d, persistent_database_connection)
key = self.generate_revision_key(url, ud, d)
rev = pd.getValue("BB_URI_HEADREVS", key)
if rev != None:
if hasattr(self, "_sortable_revision"):
return self._sortable_revision(url, ud, d)
- pd = persist_data.PersistData(d)
+ pd = persist_data.PersistData(d, persistent_database_connection)
key = self.generate_revision_key(url, ud, d)
latest_rev = self._build_revision(url, ud, d)
Why sqlite? It handles all the locking issues for us.
"""
- def __init__(self, d):
+ def __init__(self, d, persistent_database_connection):
+ if "connection" in persistent_database_connection:
+ self.cursor = persistent_database_connection["connection"].cursor()
+ return
self.cachedir = bb.data.getVar("PERSISTENT_DIR", d, True) or bb.data.getVar("CACHE", d, True)
if self.cachedir in [None, '']:
bb.msg.fatal(bb.msg.domain.PersistData, "Please set the 'PERSISTENT_DIR' or 'CACHE' variable.")
self.cachefile = os.path.join(self.cachedir, "bb_persist_data.sqlite3")
bb.msg.debug(1, bb.msg.domain.PersistData, "Using '%s' as the persistent data cache" % self.cachefile)
- self.connection = sqlite3.connect(self.cachefile, timeout=5, isolation_level=None)
+ connection = sqlite3.connect(self.cachefile, timeout=5, isolation_level=None)
+ persistent_database_connection["connection"] = connection
+ self.cursor = persistent_database_connection["connection"].cursor()
def addDomain(self, domain):
"""
Should be called before any domain is used
Creates it if it doesn't exist.
"""
- self.connection.execute("CREATE TABLE IF NOT EXISTS %s(key TEXT, value TEXT);" % domain)
+ self.cursor.execute("CREATE TABLE IF NOT EXISTS %s(key TEXT, value TEXT);" % domain)
def delDomain(self, domain):
"""
Removes a domain and all the data it contains
"""
- self.connection.execute("DROP TABLE IF EXISTS %s;" % domain)
+ self.cursor.execute("DROP TABLE IF EXISTS %s;" % domain)
def getKeyValues(self, domain):
"""
Return a list of key + value pairs for a domain
"""
ret = {}
- data = self.connection.execute("SELECT key, value from %s;" % domain)
+ data = self.cursor.execute("SELECT key, value from %s;" % domain)
for row in data:
ret[str(row[0])] = str(row[1])
"""
Return the value of a key for a domain
"""
- data = self.connection.execute("SELECT * from %s where key=?;" % domain, [key])
+ data = self.cursor.execute("SELECT * from %s where key=?;" % domain, [key])
for row in data:
return row[1]
"""
Sets the value of a key for a domain
"""
- data = self.connection.execute("SELECT * from %s where key=?;" % domain, [key])
+ data = self.cursor.execute("SELECT * from %s where key=?;" % domain, [key])
rows = 0
for row in data:
rows = rows + 1
def _execute(self, *query):
while True:
try:
- self.connection.execute(*query)
+ self.cursor.execute(*query)
return
except sqlite3.OperationalError as e:
if 'database is locked' in str(e):