summaryrefslogtreecommitdiff
path: root/lib/tdb2/tdb1_lock.c
diff options
context:
space:
mode:
authorRusty Russell <rusty@rustcorp.com.au>2011-09-14 07:37:13 +0930
committerRusty Russell <rusty@rustcorp.com.au>2011-09-14 07:37:13 +0930
commitc5b7c4aea8a1d2be1d61c3b909177cb5ef07fd62 (patch)
treedb52ef46f4f127ffba98916a42b73eff780450af /lib/tdb2/tdb1_lock.c
parent5a3f207318bee2fc2236ac1e19eda8ac2ae80829 (diff)
downloadsamba-c5b7c4aea8a1d2be1d61c3b909177cb5ef07fd62.tar.gz
samba-c5b7c4aea8a1d2be1d61c3b909177cb5ef07fd62.tar.bz2
samba-c5b7c4aea8a1d2be1d61c3b909177cb5ef07fd62.zip
tdb2: Make tdb1 use the tdb_file structure.
Because tdb2 allows multiple opens of the same TDB, we separate out the file information from the struct tdb_context. Do the same for tdb1. Signed-off-by: Rusty Russell <rusty@rustcorp.com.au> (Imported from CCAN commit 60210a73ec08a7b34ba637ad19e6749cf6dc1952)
Diffstat (limited to 'lib/tdb2/tdb1_lock.c')
-rw-r--r--lib/tdb2/tdb1_lock.c130
1 files changed, 65 insertions, 65 deletions
diff --git a/lib/tdb2/tdb1_lock.c b/lib/tdb2/tdb1_lock.c
index 0fba7423a2..85348df04d 100644
--- a/lib/tdb2/tdb1_lock.c
+++ b/lib/tdb2/tdb1_lock.c
@@ -39,9 +39,9 @@ static int fcntl_lock(struct tdb1_context *tdb,
fl.l_pid = 0;
if (waitflag)
- return fcntl(tdb->fd, F_SETLKW, &fl);
+ return fcntl(tdb->file->fd, F_SETLKW, &fl);
else
- return fcntl(tdb->fd, F_SETLK, &fl);
+ return fcntl(tdb->file->fd, F_SETLK, &fl);
}
static int fcntl_unlock(struct tdb1_context *tdb, int rw, off_t off, off_t len)
@@ -111,7 +111,7 @@ static int fcntl_unlock(struct tdb1_context *tdb, int rw, off_t off, off_t len)
fl.l_len = len;
fl.l_pid = 0;
- return fcntl(tdb->fd, F_SETLKW, &fl);
+ return fcntl(tdb->file->fd, F_SETLKW, &fl);
}
/* list -1 is the alloc list, otherwise a hash chain. */
@@ -156,7 +156,7 @@ int tdb1_brlock(struct tdb1_context *tdb,
if (!(flags & TDB_LOCK_PROBE) && errno != EAGAIN) {
tdb_logerr(tdb, TDB_ERR_LOCK, TDB_LOG_ERROR,
"tdb1_brlock failed (fd=%d) at offset %d rw_type=%d flags=%d len=%d",
- tdb->fd, offset, rw_type, flags, (int)len);
+ tdb->file->fd, offset, rw_type, flags, (int)len);
}
return -1;
}
@@ -180,7 +180,7 @@ int tdb1_brunlock(struct tdb1_context *tdb,
tdb->last_error = tdb_logerr(tdb, TDB_ERR_LOCK, TDB_LOG_ERROR,
"tdb1_brunlock failed (fd=%d) at offset"
" %d rw_type=%d len=%d",
- tdb->fd, offset, rw_type, (int)len);
+ tdb->file->fd, offset, rw_type, (int)len);
}
return ret;
}
@@ -195,15 +195,15 @@ int tdb1_allrecord_upgrade(struct tdb1_context *tdb)
{
int count = 1000;
- if (tdb->allrecord_lock.count != 1) {
+ if (tdb->file->allrecord_lock.count != 1) {
tdb->last_error = tdb_logerr(tdb, TDB_ERR_LOCK, TDB_LOG_ERROR,
"tdb1_allrecord_upgrade failed: "
"count %u too high",
- tdb->allrecord_lock.count);
+ tdb->file->allrecord_lock.count);
return -1;
}
- if (tdb->allrecord_lock.off != 1) {
+ if (tdb->file->allrecord_lock.off != 1) {
tdb->last_error = tdb_logerr(tdb, TDB_ERR_LOCK, TDB_LOG_ERROR,
"tdb1_allrecord_upgrade failed:"
" already upgraded?");
@@ -214,8 +214,8 @@ int tdb1_allrecord_upgrade(struct tdb1_context *tdb)
struct timeval tv;
if (tdb1_brlock(tdb, F_WRLCK, TDB1_FREELIST_TOP, 0,
TDB_LOCK_WAIT|TDB_LOCK_PROBE) == 0) {
- tdb->allrecord_lock.ltype = F_WRLCK;
- tdb->allrecord_lock.off = 0;
+ tdb->file->allrecord_lock.ltype = F_WRLCK;
+ tdb->file->allrecord_lock.off = 0;
return 0;
}
if (errno != EDEADLK) {
@@ -231,14 +231,14 @@ int tdb1_allrecord_upgrade(struct tdb1_context *tdb)
return -1;
}
-static struct tdb1_lock_type *tdb1_find_nestlock(struct tdb1_context *tdb,
- tdb1_off_t offset)
+static struct tdb_lock *tdb1_find_nestlock(struct tdb1_context *tdb,
+ tdb1_off_t offset)
{
unsigned int i;
- for (i=0; i<tdb->num_lockrecs; i++) {
- if (tdb->lockrecs[i].off == offset) {
- return &tdb->lockrecs[i];
+ for (i=0; i<tdb->file->num_lockrecs; i++) {
+ if (tdb->file->lockrecs[i].off == offset) {
+ return &tdb->file->lockrecs[i];
}
}
return NULL;
@@ -248,7 +248,7 @@ static struct tdb1_lock_type *tdb1_find_nestlock(struct tdb1_context *tdb,
int tdb1_nest_lock(struct tdb1_context *tdb, uint32_t offset, int ltype,
enum tdb_lock_flags flags)
{
- struct tdb1_lock_type *new_lck;
+ struct tdb_lock *new_lck;
if (offset >= lock_offset(tdb->header.hash_size)) {
tdb->last_error = tdb_logerr(tdb, TDB_ERR_LOCK, TDB_LOG_ERROR,
@@ -270,14 +270,14 @@ int tdb1_nest_lock(struct tdb1_context *tdb, uint32_t offset, int ltype,
return 0;
}
- new_lck = (struct tdb1_lock_type *)realloc(
- tdb->lockrecs,
- sizeof(*tdb->lockrecs) * (tdb->num_lockrecs+1));
+ new_lck = (struct tdb_lock *)realloc(
+ tdb->file->lockrecs,
+ sizeof(*tdb->file->lockrecs) * (tdb->file->num_lockrecs+1));
if (new_lck == NULL) {
errno = ENOMEM;
return -1;
}
- tdb->lockrecs = new_lck;
+ tdb->file->lockrecs = new_lck;
/* Since fcntl locks don't nest, we do a lock for the first one,
and simply bump the count for future ones */
@@ -285,10 +285,10 @@ int tdb1_nest_lock(struct tdb1_context *tdb, uint32_t offset, int ltype,
return -1;
}
- tdb->lockrecs[tdb->num_lockrecs].off = offset;
- tdb->lockrecs[tdb->num_lockrecs].count = 1;
- tdb->lockrecs[tdb->num_lockrecs].ltype = ltype;
- tdb->num_lockrecs++;
+ tdb->file->lockrecs[tdb->file->num_lockrecs].off = offset;
+ tdb->file->lockrecs[tdb->file->num_lockrecs].count = 1;
+ tdb->file->lockrecs[tdb->file->num_lockrecs].ltype = ltype;
+ tdb->file->num_lockrecs++;
return 0;
}
@@ -319,8 +319,8 @@ static bool have_data_locks(const struct tdb1_context *tdb)
{
unsigned int i;
- for (i = 0; i < tdb->num_lockrecs; i++) {
- if (tdb->lockrecs[i].off >= lock_offset(-1))
+ for (i = 0; i < tdb->file->num_lockrecs; i++) {
+ if (tdb->file->lockrecs[i].off >= lock_offset(-1))
return true;
}
return false;
@@ -333,12 +333,12 @@ static int tdb1_lock_list(struct tdb1_context *tdb, int list, int ltype,
bool check = false;
/* a allrecord lock allows us to avoid per chain locks */
- if (tdb->allrecord_lock.count &&
- (ltype == tdb->allrecord_lock.ltype || ltype == F_RDLCK)) {
+ if (tdb->file->allrecord_lock.count &&
+ (ltype == tdb->file->allrecord_lock.ltype || ltype == F_RDLCK)) {
return 0;
}
- if (tdb->allrecord_lock.count) {
+ if (tdb->file->allrecord_lock.count) {
tdb->last_error = TDB_ERR_LOCK;
ret = -1;
} else {
@@ -375,7 +375,7 @@ int tdb1_lock(struct tdb1_context *tdb, int list, int ltype)
int tdb1_nest_unlock(struct tdb1_context *tdb, uint32_t offset, int ltype)
{
int ret = -1;
- struct tdb1_lock_type *lck;
+ struct tdb_lock *lck;
if (tdb->flags & TDB_NOLOCK)
return 0;
@@ -413,15 +413,15 @@ int tdb1_nest_unlock(struct tdb1_context *tdb, uint32_t offset, int ltype)
* Shrink the array by overwriting the element just unlocked with the
* last array element.
*/
- *lck = tdb->lockrecs[--tdb->num_lockrecs];
+ *lck = tdb->file->lockrecs[--tdb->file->num_lockrecs];
/*
* We don't bother with realloc when the array shrinks, but if we have
* a completely idle tdb we should get rid of the locked array.
*/
- if (tdb->num_lockrecs == 0) {
- SAFE_FREE(tdb->lockrecs);
+ if (tdb->file->num_lockrecs == 0) {
+ SAFE_FREE(tdb->file->lockrecs);
}
return ret;
@@ -430,12 +430,12 @@ int tdb1_nest_unlock(struct tdb1_context *tdb, uint32_t offset, int ltype)
int tdb1_unlock(struct tdb1_context *tdb, int list, int ltype)
{
/* a global lock allows us to avoid per chain locks */
- if (tdb->allrecord_lock.count &&
- (ltype == tdb->allrecord_lock.ltype || ltype == F_RDLCK)) {
+ if (tdb->file->allrecord_lock.count &&
+ (ltype == tdb->file->allrecord_lock.ltype || ltype == F_RDLCK)) {
return 0;
}
- if (tdb->allrecord_lock.count) {
+ if (tdb->file->allrecord_lock.count) {
tdb->last_error = TDB_ERR_LOCK;
return -1;
}
@@ -470,12 +470,12 @@ static int tdb1_allrecord_check(struct tdb1_context *tdb, int ltype,
return -1;
}
- if (tdb->allrecord_lock.count && tdb->allrecord_lock.ltype == ltype) {
- tdb->allrecord_lock.count++;
+ if (tdb->file->allrecord_lock.count && tdb->file->allrecord_lock.ltype == ltype) {
+ tdb->file->allrecord_lock.count++;
return 0;
}
- if (tdb->allrecord_lock.count) {
+ if (tdb->file->allrecord_lock.count) {
/* a global lock of a different type exists */
tdb->last_error = TDB_ERR_LOCK;
return -1;
@@ -562,11 +562,11 @@ int tdb1_allrecord_lock(struct tdb1_context *tdb, int ltype,
return -1;
}
- tdb->allrecord_lock.count = 1;
+ tdb->file->allrecord_lock.count = 1;
/* If it's upgradable, it's actually exclusive so we can treat
* it as a write lock. */
- tdb->allrecord_lock.ltype = upgradable ? F_WRLCK : ltype;
- tdb->allrecord_lock.off = upgradable;
+ tdb->file->allrecord_lock.ltype = upgradable ? F_WRLCK : ltype;
+ tdb->file->allrecord_lock.off = upgradable;
if (tdb1_needs_recovery(tdb)) {
tdb1_allrecord_unlock(tdb, ltype);
@@ -590,20 +590,20 @@ int tdb1_allrecord_unlock(struct tdb1_context *tdb, int ltype)
return -1;
}
- if (tdb->allrecord_lock.count == 0) {
+ if (tdb->file->allrecord_lock.count == 0) {
tdb->last_error = TDB_ERR_LOCK;
return -1;
}
/* Upgradable locks are marked as write locks. */
- if (tdb->allrecord_lock.ltype != ltype
- && (!tdb->allrecord_lock.off || ltype != F_RDLCK)) {
+ if (tdb->file->allrecord_lock.ltype != ltype
+ && (!tdb->file->allrecord_lock.off || ltype != F_RDLCK)) {
tdb->last_error = TDB_ERR_LOCK;
return -1;
}
- if (tdb->allrecord_lock.count > 1) {
- tdb->allrecord_lock.count--;
+ if (tdb->file->allrecord_lock.count > 1) {
+ tdb->file->allrecord_lock.count--;
return 0;
}
@@ -613,8 +613,8 @@ int tdb1_allrecord_unlock(struct tdb1_context *tdb, int ltype)
return -1;
}
- tdb->allrecord_lock.count = 0;
- tdb->allrecord_lock.ltype = 0;
+ tdb->file->allrecord_lock.count = 0;
+ tdb->file->allrecord_lock.ltype = 0;
return 0;
}
@@ -671,7 +671,7 @@ int tdb1_chainunlock_read(struct tdb1_context *tdb, TDB_DATA key)
/* record lock stops delete underneath */
int tdb1_lock_record(struct tdb1_context *tdb, tdb1_off_t off)
{
- if (tdb->allrecord_lock.count) {
+ if (tdb->file->allrecord_lock.count) {
return 0;
}
return off ? tdb1_brlock(tdb, F_RDLCK, off, 1, TDB_LOCK_WAIT) : 0;
@@ -688,8 +688,8 @@ int tdb1_write_lock_record(struct tdb1_context *tdb, tdb1_off_t off)
for (i = &tdb->travlocks; i; i = i->next)
if (i->off == off)
return -1;
- if (tdb->allrecord_lock.count) {
- if (tdb->allrecord_lock.ltype == F_WRLCK) {
+ if (tdb->file->allrecord_lock.count) {
+ if (tdb->file->allrecord_lock.ltype == F_WRLCK) {
return 0;
}
return -1;
@@ -699,7 +699,7 @@ int tdb1_write_lock_record(struct tdb1_context *tdb, tdb1_off_t off)
int tdb1_write_unlock_record(struct tdb1_context *tdb, tdb1_off_t off)
{
- if (tdb->allrecord_lock.count) {
+ if (tdb->file->allrecord_lock.count) {
return 0;
}
return tdb1_brunlock(tdb, F_WRLCK, off, 1);
@@ -711,7 +711,7 @@ int tdb1_unlock_record(struct tdb1_context *tdb, tdb1_off_t off)
struct tdb1_traverse_lock *i;
uint32_t count = 0;
- if (tdb->allrecord_lock.count) {
+ if (tdb->file->allrecord_lock.count) {
return 0;
}
@@ -725,10 +725,10 @@ int tdb1_unlock_record(struct tdb1_context *tdb, tdb1_off_t off)
bool tdb1_have_extra_locks(struct tdb1_context *tdb)
{
- unsigned int extra = tdb->num_lockrecs;
+ unsigned int extra = tdb->file->num_lockrecs;
/* A transaction holds the lock for all records. */
- if (!tdb->transaction && tdb->allrecord_lock.count) {
+ if (!tdb->transaction && tdb->file->allrecord_lock.count) {
return true;
}
@@ -751,23 +751,23 @@ void tdb1_release_transaction_locks(struct tdb1_context *tdb)
{
unsigned int i, active = 0;
- if (tdb->allrecord_lock.count != 0) {
- tdb1_brunlock(tdb, tdb->allrecord_lock.ltype, TDB1_FREELIST_TOP, 0);
- tdb->allrecord_lock.count = 0;
+ if (tdb->file->allrecord_lock.count != 0) {
+ tdb1_brunlock(tdb, tdb->file->allrecord_lock.ltype, TDB1_FREELIST_TOP, 0);
+ tdb->file->allrecord_lock.count = 0;
}
- for (i=0;i<tdb->num_lockrecs;i++) {
- struct tdb1_lock_type *lck = &tdb->lockrecs[i];
+ for (i=0;i<tdb->file->num_lockrecs;i++) {
+ struct tdb_lock *lck = &tdb->file->lockrecs[i];
/* Don't release the active lock! Copy it to first entry. */
if (lck->off == TDB1_ACTIVE_LOCK) {
- tdb->lockrecs[active++] = *lck;
+ tdb->file->lockrecs[active++] = *lck;
} else {
tdb1_brunlock(tdb, lck->ltype, lck->off, 1);
}
}
- tdb->num_lockrecs = active;
- if (tdb->num_lockrecs == 0) {
- SAFE_FREE(tdb->lockrecs);
+ tdb->file->num_lockrecs = active;
+ if (tdb->file->num_lockrecs == 0) {
+ SAFE_FREE(tdb->file->lockrecs);
}
}