aboutsummaryrefslogtreecommitdiffstats
path: root/databases/evolution-data-server
diff options
context:
space:
mode:
authorkwm <kwm@FreeBSD.org>2005-03-20 21:14:03 +0800
committerkwm <kwm@FreeBSD.org>2005-03-20 21:14:03 +0800
commit5d1e41267e8182f0c75e48689532a7a212df36c0 (patch)
treebd49f8924e947cdd0c36f2e18042d3c93b9e43f5 /databases/evolution-data-server
parent222a16d2605f74596d6cae32a4aa712f3badebd6 (diff)
downloadfreebsd-ports-gnome-5d1e41267e8182f0c75e48689532a7a212df36c0.tar.gz
freebsd-ports-gnome-5d1e41267e8182f0c75e48689532a7a212df36c0.tar.zst
freebsd-ports-gnome-5d1e41267e8182f0c75e48689532a7a212df36c0.zip
Add 2 (rolled in 1) sleepycat vendor patches to the db version shipped with
evolution-data-server. Bump PORTREVISION. patch 4.1.25.1 Application, using Berkeley DB's Concurrent Data Store product with the DB_CDB_ALLDB flag set, that open databases while also holding open cursors could hang. patch 4.1.25.2 fix transaction abort hang. Obtained from: patches from sleepycat, Patch changelog from debian db 4.1 package changelog.
Diffstat (limited to 'databases/evolution-data-server')
-rw-r--r--databases/evolution-data-server/Makefile2
-rw-r--r--databases/evolution-data-server/files/patch-offical-sleepycat480
2 files changed, 481 insertions, 1 deletions
diff --git a/databases/evolution-data-server/Makefile b/databases/evolution-data-server/Makefile
index 1cbed613c81e..d2550f0af690 100644
--- a/databases/evolution-data-server/Makefile
+++ b/databases/evolution-data-server/Makefile
@@ -7,7 +7,7 @@
PORTNAME= evolution-data-server
PORTVERSION= 1.2.1
-PORTREVISION= 1
+PORTREVISION= 2
CATEGORIES= databases gnome
MASTER_SITES= ${MASTER_SITE_GNOME}
MASTER_SITE_SUBDIR= sources/${PORTNAME}/1.2
diff --git a/databases/evolution-data-server/files/patch-offical-sleepycat b/databases/evolution-data-server/files/patch-offical-sleepycat
new file mode 100644
index 000000000000..0cd4d7909728
--- /dev/null
+++ b/databases/evolution-data-server/files/patch-offical-sleepycat
@@ -0,0 +1,480 @@
+--- libdb/fileops/fop_util.c.orig Thu Nov 20 23:13:30 2003
++++ libdb/fileops/fop_util.c Fri Mar 18 20:31:10 2005
+@@ -40,7 +40,7 @@
+ u_int32_t __lockval; \
+ \
+ if (LOCKING_ON((ENV))) { \
+- __lockval = 0; \
++ __lockval = 1; \
+ __dbt.data = &__lockval; \
+ __dbt.size = sizeof(__lockval); \
+ if ((ret = (ENV)->lock_get((ENV), (ID), \
+--- libdb/dbinc/mp.h.orig Thu Nov 20 23:13:17 2003
++++ libdb/dbinc/mp.h Fri Mar 18 20:31:14 2005
+@@ -149,6 +149,13 @@
+ * region lock).
+ */
+ DB_MPOOL_STAT stat; /* Per-cache mpool statistics. */
++
++ /*
++ * We track page puts so that we can decide when allocation is never
++ * going to succeed. We don't lock the field, all we care about is
++ * if it changes.
++ */
++ u_int32_t put_counter; /* Count of page put calls. */
+ };
+
+ struct __db_mpool_hash {
+--- libdb/mp/mp_fput.c.orig Thu Nov 20 23:13:36 2003
++++ libdb/mp/mp_fput.c Fri Mar 18 20:31:14 2005
+@@ -19,6 +19,8 @@
+ #include "dbinc/db_shash.h"
+ #include "dbinc/mp.h"
+
++static void __memp_reset_lru __P((DB_ENV *, REGINFO *));
++
+ /*
+ * __memp_fput --
+ * Mpool file put function.
+@@ -198,5 +200,56 @@
+
+ MUTEX_UNLOCK(dbenv, &hp->hash_mutex);
+
++ /*
++ * On every buffer put we update the buffer generation number and check
++ * for wraparound.
++ */
++ if (++c_mp->lru_count == UINT32_T_MAX)
++ __memp_reset_lru(dbenv, dbmp->reginfo);
++
+ return (0);
++}
++
++/*
++ * __memp_reset_lru --
++ * Reset the cache LRU counter.
++ */
++static void
++__memp_reset_lru(dbenv, memreg)
++ DB_ENV *dbenv;
++ REGINFO *memreg;
++{
++ BH *bhp;
++ DB_MPOOL_HASH *hp;
++ MPOOL *c_mp;
++ int bucket;
++
++ c_mp = memreg->primary;
++
++ /*
++ * Update the counter so all future allocations will start at the
++ * bottom.
++ */
++ c_mp->lru_count -= MPOOL_BASE_DECREMENT;
++
++ /* Adjust the priority of every buffer in the system. */
++ for (hp = R_ADDR(memreg, c_mp->htab),
++ bucket = 0; bucket < c_mp->htab_buckets; ++hp, ++bucket) {
++ /*
++ * Skip empty buckets.
++ *
++ * We can check for empty buckets before locking as we
++ * only care if the pointer is zero or non-zero.
++ */
++ if (SH_TAILQ_FIRST(&hp->hash_bucket, __bh) == NULL)
++ continue;
++
++ MUTEX_LOCK(dbenv, &hp->hash_mutex);
++ for (bhp = SH_TAILQ_FIRST(&hp->hash_bucket, __bh);
++ bhp != NULL; bhp = SH_TAILQ_NEXT(bhp, hq, __bh))
++ if (bhp->priority != UINT32_T_MAX &&
++ bhp->priority > MPOOL_BASE_DECREMENT)
++ bhp->priority -= MPOOL_BASE_DECREMENT;
++ MUTEX_UNLOCK(dbenv, &hp->hash_mutex);
++ }
+ }
+--- libdb/mp/mp_alloc.c.orig Thu Nov 20 23:13:36 2003
++++ libdb/mp/mp_alloc.c Fri Mar 18 20:31:14 2005
+@@ -25,7 +25,6 @@
+ } HS;
+
+ static void __memp_bad_buffer __P((DB_MPOOL_HASH *));
+-static void __memp_reset_lru __P((DB_ENV *, REGINFO *, MPOOL *));
+
+ /*
+ * __memp_alloc --
+@@ -50,8 +49,9 @@
+ MPOOL *c_mp;
+ MPOOLFILE *bh_mfp;
+ size_t freed_space;
+- u_int32_t buckets, buffers, high_priority, max_na, priority;
+- int aggressive, ret;
++ u_int32_t buckets, buffers, high_priority, priority, put_counter;
++ u_int32_t total_buckets;
++ int aggressive, giveup, ret;
+ void *p;
+
+ dbenv = dbmp->dbenv;
+@@ -59,18 +59,13 @@
+ dbht = R_ADDR(memreg, c_mp->htab);
+ hp_end = &dbht[c_mp->htab_buckets];
+
+- buckets = buffers = 0;
+- aggressive = 0;
++ buckets = buffers = put_counter = total_buckets = 0;
++ aggressive = giveup = 0;
++ hp_tmp = NULL;
+
+ c_mp->stat.st_alloc++;
+
+ /*
+- * Get aggressive if we've tried to flush the number of pages as are
+- * in the system without finding space.
+- */
+- max_na = 5 * c_mp->htab_buckets;
+-
+- /*
+ * If we're allocating a buffer, and the one we're discarding is the
+ * same size, we don't want to waste the time to re-integrate it into
+ * the shared memory free list. If the DB_MPOOLFILE argument isn't
+@@ -81,19 +76,10 @@
+ len = (sizeof(BH) - sizeof(u_int8_t)) + mfp->stat.st_pagesize;
+
+ R_LOCK(dbenv, memreg);
+-
+- /*
+- * On every buffer allocation we update the buffer generation number
+- * and check for wraparound.
+- */
+- if (++c_mp->lru_count == UINT32_T_MAX)
+- __memp_reset_lru(dbenv, memreg, c_mp);
+-
+ /*
+ * Anything newer than 1/10th of the buffer pool is ignored during
+ * allocation (unless allocation starts failing).
+ */
+- DB_ASSERT(c_mp->lru_count > c_mp->stat.st_pages / 10);
+ high_priority = c_mp->lru_count - c_mp->stat.st_pages / 10;
+
+ /*
+@@ -120,10 +106,11 @@
+ * We're not holding the region locked here, these statistics
+ * can't be trusted.
+ */
+- if (buckets != 0) {
+- if (buckets > c_mp->stat.st_alloc_max_buckets)
+- c_mp->stat.st_alloc_max_buckets = buckets;
+- c_mp->stat.st_alloc_buckets += buckets;
++ total_buckets += buckets;
++ if (total_buckets != 0) {
++ if (total_buckets > c_mp->stat.st_alloc_max_buckets)
++ c_mp->stat.st_alloc_max_buckets = total_buckets;
++ c_mp->stat.st_alloc_buckets += total_buckets;
+ }
+ if (buffers != 0) {
+ if (buffers > c_mp->stat.st_alloc_max_pages)
+@@ -131,6 +118,12 @@
+ c_mp->stat.st_alloc_pages += buffers;
+ }
+ return (0);
++ } else if (giveup || c_mp->stat.st_pages == 0) {
++ R_UNLOCK(dbenv, memreg);
++
++ __db_err(dbenv,
++ "unable to allocate space from the buffer cache");
++ return (ret);
+ }
+
+ /*
+@@ -138,26 +131,24 @@
+ * we need. Reset our free-space counter.
+ */
+ freed_space = 0;
++ total_buckets += buckets;
++ buckets = 0;
+
+ /*
+ * Walk the hash buckets and find the next two with potentially useful
+ * buffers. Free the buffer with the lowest priority from the buckets'
+ * chains.
+ */
+- for (hp_tmp = NULL;;) {
++ for (;;) {
++ /* All pages have been freed, make one last try */
++ if (c_mp->stat.st_pages == 0)
++ goto alloc;
++
+ /* Check for wrap around. */
+ hp = &dbht[c_mp->last_checked++];
+ if (hp >= hp_end) {
+ c_mp->last_checked = 0;
+-
+- /*
+- * If we've gone through all of the hash buckets, try
+- * an allocation. If the cache is small, the old page
+- * size is small, and the new page size is large, we
+- * might have freed enough memory (but not 3 times the
+- * memory).
+- */
+- goto alloc;
++ hp = &dbht[c_mp->last_checked++];
+ }
+
+ /*
+@@ -172,39 +163,59 @@
+ /*
+ * The failure mode is when there are too many buffers we can't
+ * write or there's not enough memory in the system. We don't
+- * have a metric for deciding if allocation has no possible way
+- * to succeed, so we don't ever fail, we assume memory will be
+- * available if we wait long enough.
++ * have a way to know that allocation has no way to succeed.
++ * We fail if there were no pages returned to the cache after
++ * we've been trying for a relatively long time.
+ *
+- * Get aggressive if we've tried to flush 5 times the number of
+- * hash buckets as are in the system -- it's possible we have
+- * been repeatedly trying to flush the same buffers, although
+- * it's unlikely. Aggressive means:
++ * Get aggressive if we've tried to flush the number of hash
++ * buckets as are in the system and have not found any more
++ * space. Aggressive means:
+ *
+ * a: set a flag to attempt to flush high priority buffers as
+ * well as other buffers.
+ * b: sync the mpool to force out queue extent pages. While we
+ * might not have enough space for what we want and flushing
+ * is expensive, why not?
+- * c: sleep for a second -- hopefully someone else will run and
+- * free up some memory. Try to allocate memory too, in case
+- * the other thread returns its memory to the region.
+- * d: look at a buffer in every hash bucket rather than choose
++ * c: look at a buffer in every hash bucket rather than choose
+ * the more preferable of two.
++ * d: start to think about giving up.
++ *
++ * If we get here twice, sleep for a second, hopefully someone
++ * else will run and free up some memory.
++ *
++ * Always try to allocate memory too, in case some other thread
++ * returns its memory to the region.
+ *
+ * !!!
+ * This test ignores pathological cases like no buffers in the
+ * system -- that shouldn't be possible.
+ */
+- if ((++buckets % max_na) == 0) {
+- aggressive = 1;
+-
++ if ((++buckets % c_mp->htab_buckets) == 0) {
++ if (freed_space > 0)
++ goto alloc;
+ R_UNLOCK(dbenv, memreg);
+
+- (void)__memp_sync_int(
+- dbenv, NULL, 0, DB_SYNC_ALLOC, NULL);
+-
+- (void)__os_sleep(dbenv, 1, 0);
++ switch (++aggressive) {
++ case 1:
++ break;
++ case 2:
++ put_counter = c_mp->put_counter;
++ /* FALLTHROUGH */
++ case 3:
++ case 4:
++ case 5:
++ case 6:
++ (void)__memp_sync_int(
++ dbenv, NULL, 0, DB_SYNC_ALLOC, NULL);
++
++ (void)__os_sleep(dbenv, 1, 0);
++ break;
++ default:
++ aggressive = 1;
++ if (put_counter == c_mp->put_counter)
++ giveup = 1;
++ break;
++ }
+
+ R_LOCK(dbenv, memreg);
+ goto alloc;
+@@ -277,7 +288,8 @@
+ * thread may have acquired this buffer and incremented the ref
+ * count after we wrote it, in which case we can't have it.
+ *
+- * If there's a write error, avoid selecting this buffer again
++ * If there's a write error and we're having problems finding
++ * something to allocate, avoid selecting this buffer again
+ * by making it the bucket's least-desirable buffer.
+ */
+ if (ret != 0 || bhp->ref != 0) {
+@@ -301,6 +313,8 @@
+
+ freed_space += __db_shsizeof(bhp);
+ __memp_bhfree(dbmp, hp, bhp, 1);
++ if (aggressive > 1)
++ aggressive = 1;
+
+ /*
+ * Unlock this hash bucket and re-acquire the region lock. If
+@@ -360,54 +374,6 @@
+
+ /* Reset the hash bucket's priority. */
+ hp->hash_priority = SH_TAILQ_FIRST(&hp->hash_bucket, __bh)->priority;
+-}
+-
+-/*
+- * __memp_reset_lru --
+- * Reset the cache LRU counter.
+- */
+-static void
+-__memp_reset_lru(dbenv, memreg, c_mp)
+- DB_ENV *dbenv;
+- REGINFO *memreg;
+- MPOOL *c_mp;
+-{
+- BH *bhp;
+- DB_MPOOL_HASH *hp;
+- int bucket;
+-
+- /*
+- * Update the counter so all future allocations will start at the
+- * bottom.
+- */
+- c_mp->lru_count -= MPOOL_BASE_DECREMENT;
+-
+- /* Release the region lock. */
+- R_UNLOCK(dbenv, memreg);
+-
+- /* Adjust the priority of every buffer in the system. */
+- for (hp = R_ADDR(memreg, c_mp->htab),
+- bucket = 0; bucket < c_mp->htab_buckets; ++hp, ++bucket) {
+- /*
+- * Skip empty buckets.
+- *
+- * We can check for empty buckets before locking as we
+- * only care if the pointer is zero or non-zero.
+- */
+- if (SH_TAILQ_FIRST(&hp->hash_bucket, __bh) == NULL)
+- continue;
+-
+- MUTEX_LOCK(dbenv, &hp->hash_mutex);
+- for (bhp = SH_TAILQ_FIRST(&hp->hash_bucket, __bh);
+- bhp != NULL; bhp = SH_TAILQ_NEXT(bhp, hq, __bh))
+- if (bhp->priority != UINT32_T_MAX &&
+- bhp->priority > MPOOL_BASE_DECREMENT)
+- bhp->priority -= MPOOL_BASE_DECREMENT;
+- MUTEX_UNLOCK(dbenv, &hp->hash_mutex);
+- }
+-
+- /* Reacquire the region lock. */
+- R_LOCK(dbenv, memreg);
+ }
+
+ #ifdef DIAGNOSTIC
+--- libdb/dbreg/dbreg_rec.c.orig Thu Nov 20 23:13:19 2003
++++ libdb/dbreg/dbreg_rec.c Fri Mar 18 20:31:14 2005
+@@ -174,19 +174,20 @@
+ * Typically, closes should match an open which means
+ * that if this is a close, there should be a valid
+ * entry in the dbentry table when we get here,
+- * however there is an exception. If this is an
++ * however there are exceptions. 1. If this is an
+ * OPENFILES pass, then we may have started from
+ * a log file other than the first, and the
+ * corresponding open appears in an earlier file.
+- * We can ignore that case, but all others are errors.
++ * 2. If we are undoing an open on an abort or
++ * recovery, it's possible that we failed after
++ * the log record, but before we actually entered
++ * a handle here.
+ */
+ dbe = &dblp->dbentry[argp->fileid];
+ if (dbe->dbp == NULL && !dbe->deleted) {
+ /* No valid entry here. */
+- if ((argp->opcode != LOG_CLOSE &&
+- argp->opcode != LOG_RCLOSE) ||
+- (op != DB_TXN_OPENFILES &&
+- op !=DB_TXN_POPENFILES)) {
++ if (DB_REDO(op) ||
++ argp->opcode == LOG_CHECKPOINT) {
+ __db_err(dbenv,
+ "Improper file close at %lu/%lu",
+ (u_long)lsnp->file,
+--- libdb/env/env_recover.c.orig Thu Nov 20 23:13:20 2003
++++ libdb/env/env_recover.c Fri Mar 18 20:31:14 2005
+@@ -232,12 +232,9 @@
+ * we'll still need to do a vtruncate based on information we haven't
+ * yet collected.
+ */
+- if (ret == DB_NOTFOUND) {
++ if (ret == DB_NOTFOUND)
+ ret = 0;
+- if (max_lsn == NULL)
+- goto done;
+- }
+- if (ret != 0)
++ else if (ret != 0)
+ goto err;
+
+ hi_txn = txnid;
+@@ -331,7 +328,7 @@
+
+ /* Find a low txnid. */
+ ret = 0;
+- do {
++ if (hi_txn != 0) do {
+ /* txnid is after rectype, which is a u_int32. */
+ memcpy(&txnid,
+ (u_int8_t *)data.data + sizeof(u_int32_t), sizeof(txnid));
+@@ -344,11 +341,8 @@
+ * There are no transactions and we're not recovering to an LSN (see
+ * above), so there is nothing to do.
+ */
+- if (ret == DB_NOTFOUND) {
++ if (ret == DB_NOTFOUND)
+ ret = 0;
+- if (max_lsn == NULL)
+- goto done;
+- }
+
+ /* Reset to the first lsn. */
+ if (ret != 0 || (ret = logc->get(logc, &first_lsn, &data, DB_SET)) != 0)
+@@ -367,6 +361,10 @@
+ txninfo, &data, &first_lsn, &last_lsn, nfiles, 1)) != 0)
+ goto err;
+
++ /* If there were no transactions, then we can bail out early. */
++ if (hi_txn == 0 && max_lsn == NULL)
++ goto done;
++
+ /*
+ * Pass #2.
+ *
+@@ -483,6 +481,7 @@
+ if ((ret = __dbreg_close_files(dbenv)) != 0)
+ goto err;
+
++done:
+ if (max_lsn != NULL) {
+ region->last_ckp = ((DB_TXNHEAD *)txninfo)->ckplsn;
+
+@@ -538,7 +537,8 @@
+ __db_err(dbenv, "Recovery complete at %.24s", ctime(&now));
+ __db_err(dbenv, "%s %lx %s [%lu][%lu]",
+ "Maximum transaction ID",
+- ((DB_TXNHEAD *)txninfo)->maxid,
++ txninfo == NULL ? TXN_MINIMUM :
++ ((DB_TXNHEAD *)txninfo)->maxid,
+ "Recovery checkpoint",
+ (u_long)region->last_ckp.file,
+ (u_long)region->last_ckp.offset);
+@@ -550,7 +550,6 @@
+ (u_long)lsn.file, (u_long)lsn.offset, pass);
+ }
+
+-done:
+ err: if (lockid != DB_LOCK_INVALIDID) {
+ if ((t_ret = __rep_unlockpages(dbenv, lockid)) != 0 && ret == 0)
+ ret = t_ret;