summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorJeremy Allison <jra@samba.org>2000-05-25 20:05:50 +0000
committerJeremy Allison <jra@samba.org>2000-05-25 20:05:50 +0000
commit9c4079d8c3b158369585d33f7b19fc7bd585401e (patch)
tree2cdf4693e78378ed7ffed054037d1e829ba6d909
parentd29f8af77babe826158191a1caaea3846626236a (diff)
downloadsamba-9c4079d8c3b158369585d33f7b19fc7bd585401e.tar.gz
samba-9c4079d8c3b158369585d33f7b19fc7bd585401e.tar.bz2
samba-9c4079d8c3b158369585d33f7b19fc7bd585401e.zip
Fixed misunderstanding found during CIFS conference. Overlapping lock
ranges (not just included lock ranges) should be handled correctly. UNIT test still needed. Jeremy. (This used to be commit 07872298e3ee8b4b50b69cb4e49b88635792128e)
-rw-r--r--source3/locking/posix.c618
1 files changed, 278 insertions, 340 deletions
diff --git a/source3/locking/posix.c b/source3/locking/posix.c
index d1edb1ef57..6ab4bef4cc 100644
--- a/source3/locking/posix.c
+++ b/source3/locking/posix.c
@@ -345,65 +345,22 @@ static BOOL delete_posix_lock_entry_by_index(files_struct *fsp, size_t entry)
}
/****************************************************************************
- Add an entry into the POSIX locking tdb. Returns the number of records that
- completely overlap this request, or -1 on error. entry_num gets set to the
- index number of the added lock (used in case we need to delete *exactly*
- this entry).
+ Add an entry into the POSIX locking tdb. We return the index number of the
+ added lock (used in case we need to delete *exactly* this entry). Returns
+ False on fail, True on success.
****************************************************************************/
-static int add_posix_lock_entry(files_struct *fsp, SMB_OFF_T start, SMB_OFF_T size, int lock_type, size_t *entry_num)
+static BOOL add_posix_lock_entry(files_struct *fsp, SMB_OFF_T start, SMB_OFF_T size, int lock_type, size_t *pentry_num)
{
TDB_DATA kbuf = locking_key_fsp(fsp);
TDB_DATA dbuf;
struct posix_lock pl;
- struct posix_lock *entries;
- size_t i, count;
- int num_overlapping_records = 0;
- *entry_num = 0;
-
- /*
- * Windows is very strange. It allows read locks to be overlayed on
- * a write lock, but leaves the write lock in force until the first
- * unlock. It also reference counts the locks. This means the following sequence :
- *
- * process1 process2
- * ------------------------------------------------------------------------
- * WRITE LOCK : start = 0, len = 10
- * READ LOCK: start =0, len = 10 - FAIL
- * READ LOCK : start = 5, len = 2
- * READ LOCK: start =0, len = 10 - FAIL
- * UNLOCK : start = 0, len = 10
- * READ LOCK: start =0, len = 10 - OK
- *
- * Under POSIX, the same sequence in steps 1 and 2 would not be reference counted, but
- * would leave a single read lock over the 0-10 region. In order to
- * re-create Windows semantics mapped to POSIX locks, we create multiple TDB
- * entries, one for each overlayed lock request. We are guarenteed by the brlock
- * semantics that if a write lock is added, then it will be first in the array.
- */
-
dbuf.dptr = NULL;
dbuf = tdb_fetch(posix_lock_tdb, kbuf);
- count = (size_t)(dbuf.dsize / sizeof(pl));
- entries = (struct posix_lock *)dbuf.dptr;
-
- /*
- * Ensure we look for overlapping entries *before*
- * we add this entry. Count the number of entries
- * that completely overlap this request.
- */
-
- for (i = 0; i < count; i++) {
- struct posix_lock *entry = &entries[i];
-
- if (fsp->fd == entry->fd &&
- start >= entry->start &&
- start + size <= entry->start + entry->size)
- num_overlapping_records++;
- }
+ *pentry_num = (size_t)(dbuf.dsize / sizeof(pl));
/*
* Add new record.
@@ -423,8 +380,6 @@ static int add_posix_lock_entry(files_struct *fsp, SMB_OFF_T start, SMB_OFF_T si
memcpy(dbuf.dptr + dbuf.dsize, &pl, sizeof(pl));
dbuf.dsize += sizeof(pl);
- *entry_num = count;
-
if (tdb_store(posix_lock_tdb, kbuf, dbuf, TDB_REPLACE) == -1) {
DEBUG(0,("add_posix_lock: Failed to add lock entry on file %s\n", fsp->fsp_name));
goto fail;
@@ -432,23 +387,36 @@ static int add_posix_lock_entry(files_struct *fsp, SMB_OFF_T start, SMB_OFF_T si
free(dbuf.dptr);
- DEBUG(10,("add_posix_lock: File %s: type = %s: start=%.0f size=%.0f: num_records = %d : dev=%.0f inode=%.0f\n",
- fsp->fsp_name, posix_lock_type_name(lock_type), (double)start, (double)size, num_overlapping_records,
+ DEBUG(10,("add_posix_lock: File %s: type = %s: start=%.0f size=%.0f: dev=%.0f inode=%.0f\n",
+ fsp->fsp_name, posix_lock_type_name(lock_type), (double)start, (double)size,
(double)fsp->dev, (double)fsp->inode ));
- return num_overlapping_records;
+ return True;
fail:
if (dbuf.dptr)
free(dbuf.dptr);
- *entry_num = 0;
- return -1;
+ return False;
+}
+
+/****************************************************************************
+ Calculate if locks have any overlap at all.
+****************************************************************************/
+
+static BOOL does_lock_overlap(SMB_OFF_T start1, SMB_OFF_T size1, SMB_OFF_T start2, SMB_OFF_T size2)
+{
+ if (start1 >= start2 && start1 <= start2 + size2)
+ return True;
+
+ if (start1 < start2 && start1 + size1 > start2);
+ return True;
+
+ return False;
}
/****************************************************************************
Delete an entry from the POSIX locking tdb. Returns a copy of the entry being
- deleted and the number of records that are completely overlapped by this one,
- or -1 on error.
+ deleted and the number of records that are overlapped by this one, or -1 on error.
****************************************************************************/
static int delete_posix_lock_entry(files_struct *fsp, SMB_OFF_T start, SMB_OFF_T size, struct posix_lock *pl)
@@ -510,23 +478,20 @@ static int delete_posix_lock_entry(files_struct *fsp, SMB_OFF_T start, SMB_OFF_T
/*
* Count the number of entries that are
- * overlapped completely by this unlock request.
- * (Note that this is the reverse of the test in
- * add_posix_lock).
+ * overlapped by this unlock request.
*/
for (i = 0; i < count; i++) {
struct posix_lock *entry = &locks[i];
if (fsp->fd == entry->fd &&
- entry->start >= start &&
- entry->start + entry->size <= start + size)
+ does_lock_overlap( start, size, entry->start, entry->size))
num_overlapping_records++;
}
- DEBUG(10,("delete_posix_lock_entry: type = %s: start=%.0f size=%.0f, num_records = %d\n",
- posix_lock_type_name(pl->lock_type), (double)pl->start, (double)pl->size,
- (unsigned int)num_overlapping_records ));
+ DEBUG(10,("delete_posix_lock_entry: type = %s: start=%.0f size=%.0f, num_records = %d\n",
+ posix_lock_type_name(pl->lock_type), (double)pl->start, (double)pl->size,
+ (unsigned int)num_overlapping_records ));
if (dbuf.dptr)
free(dbuf.dptr);
@@ -579,177 +544,81 @@ static int map_posix_lock_type( files_struct *fsp, enum brl_type lock_type)
static BOOL posix_lock_in_range(SMB_OFF_T *offset_out, SMB_OFF_T *count_out,
SMB_BIG_UINT u_offset, SMB_BIG_UINT u_count)
{
- SMB_OFF_T offset;
- SMB_OFF_T count;
-
-#if defined(LARGE_SMB_OFF_T) && !defined(HAVE_BROKEN_FCNTL64_LOCKS)
-
- SMB_OFF_T mask2 = ((SMB_OFF_T)0x4) << (SMB_OFF_T_BITS-4);
- SMB_OFF_T mask = (mask2<<1);
- SMB_OFF_T neg_mask = ~mask;
-
- /*
- * In this case SMB_OFF_T is 64 bits,
- * and the underlying system can handle 64 bit signed locks.
- * Cast to signed type.
- */
-
- offset = (SMB_OFF_T)u_offset;
- count = (SMB_OFF_T)u_count;
+ SMB_OFF_T offset = (SMB_OFF_T)u_offset;
+ SMB_OFF_T count = (SMB_OFF_T)u_count;
/*
- * Deal with a very common case of count of all ones.
- * (lock entire file).
+ * For the type of system we are, attempt to
+ * find the maximum positive lock offset as an SMB_OFF_T.
*/
- if(count == (SMB_OFF_T)-1)
- count &= ~mask;
+#if defined(LARGE_SMB_OFF_T) && !defined(HAVE_BROKEN_FCNTL64_LOCKS)
/*
- * POSIX lock ranges cannot be negative.
- * Fail if any combination becomes negative.
+ * In this case SMB_OFF_T is 64 bits,
+ * and the underlying system can handle 64 bit signed locks.
*/
- if(offset < 0 || count < 0 || (offset + count < 0)) {
- DEBUG(10,("posix_lock_in_range: negative range: offset = %.0f, count = %.0f. Ignoring lock.\n",
- (double)offset, (double)count ));
- return False;
- }
-
- /*
- * In this case SMB_OFF_T is 64 bits, the offset and count
- * fit within the positive range, and the underlying
- * system can handle 64 bit locks. Just return as the
- * cast values are ok.
- */
+ SMB_OFF_T mask2 = ((SMB_OFF_T)0x4) << (SMB_OFF_T_BITS-4);
+ SMB_OFF_T mask = (mask2<<1);
+ SMB_OFF_T max_positive_lock_offset = ~mask;
#else /* !LARGE_SMB_OFF_T || HAVE_BROKEN_FCNTL64_LOCKS */
/*
* In this case either SMB_OFF_T is 32 bits,
* or the underlying system cannot handle 64 bit signed locks.
- * Either way we have to try and mangle to fit within 31 bits.
- * This is difficult.
+ * All offsets & counts must be 2^31 or less.
*/
-#if defined(HAVE_BROKEN_FCNTL64_LOCKS)
+ SMB_OFF_T max_positive_lock_offset = 0x7FFFFFFF;
- /*
- * SMB_OFF_T is 64 bits, but we need to use 31 bits due to
- * broken large locking.
- */
+#endif /* !LARGE_SMB_OFF_T || HAVE_BROKEN_FCNTL64_LOCKS */
/*
- * Deal with a very common case of count of all ones.
- * (lock entire file).
+ * If the given offset was > max_positive_lock_offset then we cannot map this at all
+ * ignore this lock.
*/
- if(u_count == (SMB_BIG_UINT)-1)
- count = 0x7FFFFFFF;
-
- if(((u_offset >> 32) & 0xFFFFFFFF) || ((u_count >> 32) & 0xFFFFFFFF)) {
- DEBUG(10,("posix_lock_in_range: top 32 bits not zero. offset = %.0f, count = %.0f. Ignoring lock.\n",
- (double)u_offset, (double)u_count ));
- /* Top 32 bits of offset or count were not zero. */
+ if (u_offset & ~((SMB_BIG_UINT)max_positive_lock_offset)) {
+ DEBUG(10,("posix_lock_in_range: (offset = %.0f) offset > %.0f and we cannot handle this. Ignoring lock.\n",
+ (double)u_offset, (double)((SMB_BIG_UINT)max_positive_lock_offset) ));
return False;
}
- /* Cast from 64 bits unsigned to 64 bits signed. */
- offset = (SMB_OFF_T)u_offset;
- count = (SMB_OFF_T)u_count;
-
- /*
- * Check if we are within the 2^31 range.
- */
-
- {
- int32 low_offset = (int32)offset;
- int32 low_count = (int32)count;
-
- if(low_offset < 0 || low_count < 0 || (low_offset + low_count < 0)) {
- DEBUG(10,("posix_lock_in_range: not within 2^31 range. low_offset = %d, low_count = %d. Ignoring lock.\n",
- low_offset, low_count ));
- return False;
- }
- }
-
- /*
- * Ok - we can map from a 64 bit number to a 31 bit lock.
- */
-
-#else /* HAVE_BROKEN_FCNTL64_LOCKS */
-
/*
- * SMB_OFF_T is 32 bits.
+ * We must truncate the offset and count to less than max_positive_lock_offset.
*/
-#if defined(HAVE_LONGLONG)
+ offset &= max_positive_lock_offset;
+ count &= max_positive_lock_offset;
- /*
- * SMB_BIG_UINT is 64 bits, we can do a 32 bit shift.
- */
/*
* Deal with a very common case of count of all ones.
* (lock entire file).
*/
- if(u_count == (SMB_BIG_UINT)-1)
- count = 0x7FFFFFFF;
-
- if(((u_offset >> 32) & 0xFFFFFFFF) || ((u_count >> 32) & 0xFFFFFFFF)) {
- DEBUG(10,("posix_lock_in_range: top 32 bits not zero. u_offset = %.0f, u_count = %.0f. Ignoring lock.\n",
- (double)u_offset, (double)u_count ));
- return False;
- }
-
- /* Cast from 64 bits unsigned to 32 bits signed. */
- offset = (SMB_OFF_T)u_offset;
- count = (SMB_OFF_T)u_count;
-
- /*
- * Check if we are within the 2^31 range.
- */
-
- if(offset < 0 || count < 0 || (offset + count < 0)) {
- DEBUG(10,("posix_lock_in_range: not within 2^31 range. offset = %d, count = %d. Ignoring lock.\n",
- (int)offset, (int)count ));
- return False;
- }
-
-#else /* HAVE_LONGLONG */
-
- /*
- * SMB_BIG_UINT and SMB_OFF_T are both 32 bits,
- * just cast.
- */
+ if(count == (SMB_OFF_T)-1)
+ count = max_positive_lock_offset;
/*
- * Deal with a very common case of count of all ones.
- * (lock entire file).
+ * Truncate count to end at max lock offset.
*/
- if(u_count == (SMB_BIG_UINT)-1)
- count = 0x7FFFFFFF;
-
- /* Cast from 32 bits unsigned to 32 bits signed. */
- offset = (SMB_OFF_T)u_offset;
- count = (SMB_OFF_T)u_count;
+ if (offset + count < 0 || offset + count > max_positive_lock_offset)
+ count = max_positive_lock_offset - offset;
/*
- * Check if we are within the 2^31 range.
+ * If we ate all the count, ignore this lock.
*/
- if(offset < 0 || count < 0 || (offset + count < 0)) {
- DEBUG(10,("posix_lock_in_range: not within 2^31 range. offset = %d, count = %d. Ignoring lock.\n",
- (int)offset, (int)count ));
+ if (count == 0) {
+ DEBUG(10,("posix_lock_in_range: Count = 0. Ignoring lock u_offset = %.0f, u_count = %.0f\n",
+ (double)u_offset, (double)u_count ));
return False;
}
-#endif /* HAVE_LONGLONG */
-#endif /* LARGE_SMB_OFF_T */
-#endif /* !LARGE_SMB_OFF_T || HAVE_BROKEN_FCNTL64_LOCKS */
-
/*
* The mapping was successful.
*/
@@ -914,89 +783,25 @@ BOOL is_posix_locked(files_struct *fsp, SMB_BIG_UINT u_offset, SMB_BIG_UINT u_co
return posix_fcntl_lock(fsp,SMB_F_GETLK,offset,count,posix_lock_type);
}
-/****************************************************************************
- POSIX function to acquire a lock. Returns True if the
- lock could be granted, False if not.
-****************************************************************************/
-
-BOOL set_posix_lock(files_struct *fsp, SMB_BIG_UINT u_offset, SMB_BIG_UINT u_count, enum brl_type lock_type)
-{
- SMB_OFF_T offset;
- SMB_OFF_T count;
- BOOL ret = True;
- size_t entry_num = 0;
- int posix_lock_type = map_posix_lock_type(fsp,lock_type);
- int num_overlapping_records;
-
- DEBUG(5,("set_posix_lock: File %s, offset = %.0f, count = %.0f, type = %s\n",
- fsp->fsp_name, (double)u_offset, (double)u_count, posix_lock_type_name(lock_type) ));
-
- /*
- * If the requested lock won't fit in the POSIX range, we will
- * pretend it was successful.
- */
-
- if(!posix_lock_in_range(&offset, &count, u_offset, u_count))
- return True;
-
- /*
- * Note that setting multiple overlapping locks on different
- * file descriptors will not be held separately by the kernel (POSIX
- * braindamage), but will be merged into one continuous lock
- * range. We cope with this case in the release_posix_lock code
- * below. We need to add the posix lock entry into the tdb before
- * doing the real posix lock call to deal with the locking overlay
- * case described above in add_posix_lock_entry().
- */
-
- num_overlapping_records = add_posix_lock_entry(fsp,offset,count,posix_lock_type,&entry_num);
-
- if (num_overlapping_records == -1) {
- DEBUG(0,("set_posix_lock: Unable to create posix lock entry !\n"));
- return False;
- }
-
- /*
- * num_overlapping_records is the count of lock records that
- * completely contain this request on the same fsp. Only bother
- * with the real lock request if there are none, otherwise ignore.
- */
-
- if (num_overlapping_records == 0) {
- /*
- * First lock entry for this range created. Do a real POSIX lock.
- */
- ret = posix_fcntl_lock(fsp,SMB_F_SETLK,offset,count,posix_lock_type);
-
- /*
- * Oops, POSIX lock failed, delete the tdb entry we just added.
- */
- if (!ret)
- delete_posix_lock_entry_by_index(fsp,entry_num);
- }
-
- return ret;
-}
-
/*
* Structure used when splitting a lock range
* into a POSIX lock range. Doubly linked list.
*/
-struct unlock_list {
- struct unlock_list *next;
- struct unlock_list *prev;
+struct lock_list {
+ struct lock_list *next;
+ struct lock_list *prev;
SMB_OFF_T start;
SMB_OFF_T size;
};
/****************************************************************************
Create a list of lock ranges that don't overlap a given range. Used in calculating
- POSIX lock unlocks. This is a difficult function that requires ASCII art to
+ POSIX locks and unlocks. This is a difficult function that requires ASCII art to
understand it :-).
****************************************************************************/
-static struct unlock_list *posix_unlock_list(TALLOC_CTX *ctx, struct unlock_list *ulhead, files_struct *fsp)
+static struct lock_list *posix_lock_list(TALLOC_CTX *ctx, struct lock_list *lhead, files_struct *fsp)
{
TDB_DATA kbuf = locking_key_fsp(fsp);
TDB_DATA dbuf;
@@ -1007,9 +812,8 @@ static struct unlock_list *posix_unlock_list(TALLOC_CTX *ctx, struct unlock_list
dbuf = tdb_fetch(posix_lock_tdb, kbuf);
- if (!dbuf.dptr) {
- return ulhead;
- }
+ if (!dbuf.dptr)
+ return lhead;
locks = (struct posix_lock *)dbuf.dptr;
num_locks = (size_t)(dbuf.dsize / sizeof(*locks));
@@ -1019,48 +823,48 @@ static struct unlock_list *posix_unlock_list(TALLOC_CTX *ctx, struct unlock_list
* Quit if the list is deleted.
*/
- DEBUG(10,("posix_unlock_list: curr: start=%.0f,size=%.0f\n",
- (double)ulhead->start, (double)ulhead->size ));
+ DEBUG(10,("posix_lock_list: curr: start=%.0f,size=%.0f\n",
+ (double)lhead->start, (double)lhead->size ));
- for (i=0; i<num_locks && ulhead; i++) {
+ for (i=0; i<num_locks && lhead; i++) {
struct posix_lock *lock = &locks[i];
- struct unlock_list *ul_curr;
+ struct lock_list *l_curr;
/*
- * Walk the unlock list, checking for overlaps. Note that
- * the unlock list can expand within this loop if the current
+ * Walk the lock list, checking for overlaps. Note that
+ * the lock list can expand within this loop if the current
* range being examined needs to be split.
*/
- for (ul_curr = ulhead; ul_curr;) {
+ for (l_curr = lhead; l_curr;) {
- DEBUG(10,("posix_unlock_list: lock: fd=%d: start=%.0f,size=%.0f:type=%s", lock->fd,
+ DEBUG(10,("posix_lock_list: lock: fd=%d: start=%.0f,size=%.0f:type=%s", lock->fd,
(double)lock->start, (double)lock->size, posix_lock_type_name(lock->lock_type) ));
- if ( (ul_curr->start >= (lock->start + lock->size)) ||
- (lock->start >= (ul_curr->start + ul_curr->size))) {
+ if ( (l_curr->start >= (lock->start + lock->size)) ||
+ (lock->start >= (l_curr->start + l_curr->size))) {
/* No overlap with this lock - leave this range alone. */
/*********************************************
+---------+
- | ul_curr |
+ | l_curr |
+---------+
+-------+
| lock |
+-------+
OR....
+---------+
- | ul_curr |
+ | l_curr |
+---------+
**********************************************/
DEBUG(10,("no overlap case.\n" ));
- ul_curr = ul_curr->next;
+ l_curr = l_curr->next;
- } else if ( (ul_curr->start >= lock->start) &&
- (ul_curr->start + ul_curr->size <= lock->start + lock->size) ) {
+ } else if ( (l_curr->start >= lock->start) &&
+ (l_curr->start + l_curr->size <= lock->start + lock->size) ) {
/*
* This unlock is completely overlapped by this existing lock range
@@ -1068,26 +872,26 @@ OR....
*/
/*********************************************
+---------+
- | ul_curr |
+ | l_curr |
+---------+
+---------------------------+
| lock |
+---------------------------+
**********************************************/
/* Save the next pointer */
- struct unlock_list *ul_next = ul_curr->next;
+ struct lock_list *ul_next = l_curr->next;
DEBUG(10,("delete case.\n" ));
- DLIST_REMOVE(ulhead, ul_curr);
- if(ulhead == NULL)
+ DLIST_REMOVE(lhead, l_curr);
+ if(lhead == NULL)
break; /* No more list... */
- ul_curr = ul_next;
+ l_curr = ul_next;
- } else if ( (ul_curr->start >= lock->start) &&
- (ul_curr->start < lock->start + lock->size) &&
- (ul_curr->start + ul_curr->size > lock->start + lock->size) ) {
+ } else if ( (l_curr->start >= lock->start) &&
+ (l_curr->start < lock->start + lock->size) &&
+ (l_curr->start + l_curr->size > lock->start + lock->size) ) {
/*
* This unlock overlaps the existing lock range at the high end.
@@ -1095,28 +899,28 @@ OR....
*/
/*********************************************
+---------------+
- | ul_curr |
+ | l_curr |
+---------------+
+---------------+
| lock |
+---------------+
BECOMES....
+-------+
- |ul_curr|
+ | l_curr|
+-------+
**********************************************/
- ul_curr->size = (ul_curr->start + ul_curr->size) - (lock->start + lock->size);
- ul_curr->start = lock->start + lock->size;
+ l_curr->size = (l_curr->start + l_curr->size) - (lock->start + lock->size);
+ l_curr->start = lock->start + lock->size;
DEBUG(10,("truncate high case: start=%.0f,size=%.0f\n",
- (double)ul_curr->start, (double)ul_curr->size ));
+ (double)l_curr->start, (double)l_curr->size ));
- ul_curr = ul_curr->next;
+ l_curr = l_curr->next;
- } else if ( (ul_curr->start < lock->start) &&
- (ul_curr->start + ul_curr->size > lock->start) &&
- (ul_curr->start + ul_curr->size <= lock->start + lock->size) ) {
+ } else if ( (l_curr->start < lock->start) &&
+ (l_curr->start + l_curr->size > lock->start) &&
+ (l_curr->start + l_curr->size <= lock->start + lock->size) ) {
/*
* This unlock overlaps the existing lock range at the low end.
@@ -1124,26 +928,26 @@ BECOMES....
*/
/*********************************************
+---------------+
- | ul_curr |
+ | l_curr |
+---------------+
+---------------+
| lock |
+---------------+
BECOMES....
+-------+
- |ul_curr|
+ | l_curr|
+-------+
**********************************************/
- ul_curr->size = lock->start - ul_curr->start;
+ l_curr->size = lock->start - l_curr->start;
DEBUG(10,("truncate low case: start=%.0f,size=%.0f\n",
- (double)ul_curr->start, (double)ul_curr->size ));
+ (double)l_curr->start, (double)l_curr->size ));
- ul_curr = ul_curr->next;
+ l_curr = l_curr->next;
- } else if ( (ul_curr->start < lock->start) &&
- (ul_curr->start + ul_curr->size > lock->start + lock->size) ) {
+ } else if ( (l_curr->start < lock->start) &&
+ (l_curr->start + l_curr->size > lock->start + lock->size) ) {
/*
* Worst case scenario. Unlock request completely overlaps an existing
* lock range. Split the request into two, push the new (upper) request
@@ -1152,46 +956,46 @@ BECOMES....
*/
/*********************************************
+---------------------------+
- | ul_curr |
+ | l_curr |
+---------------------------+
+---------+
| lock |
+---------+
BECOMES.....
+-------+ +---------+
- |ul_curr| |ul_new |
+ | l_curr| | l_new |
+-------+ +---------+
**********************************************/
- struct unlock_list *ul_new = (struct unlock_list *)talloc(ctx,
- sizeof(struct unlock_list));
+ struct lock_list *l_new = (struct lock_list *)talloc(ctx,
+ sizeof(struct lock_list));
- if(ul_new == NULL) {
- DEBUG(0,("posix_unlock_list: talloc fail.\n"));
+ if(l_new == NULL) {
+ DEBUG(0,("posix_lock_list: talloc fail.\n"));
return NULL; /* The talloc_destroy takes care of cleanup. */
}
- ZERO_STRUCTP(ul_new);
- ul_new->start = lock->start + lock->size;
- ul_new->size = ul_curr->start + ul_curr->size - ul_new->start;
+ ZERO_STRUCTP(l_new);
+ l_new->start = lock->start + lock->size;
+ l_new->size = l_curr->start + l_curr->size - l_new->start;
- /* Truncate the ul_curr. */
- ul_curr->size = lock->start - ul_curr->start;
+ /* Truncate the l_curr. */
+ l_curr->size = lock->start - l_curr->start;
DEBUG(10,("split case: curr: start=%.0f,size=%.0f \
-new: start=%.0f,size=%.0f\n", (double)ul_curr->start, (double)ul_curr->size,
- (double)ul_new->start, (double)ul_new->size ));
+new: start=%.0f,size=%.0f\n", (double)l_curr->start, (double)l_curr->size,
+ (double)l_new->start, (double)l_new->size ));
/*
- * Add into the dlink list after the ul_curr point - NOT at ulhead.
+ * Add into the dlink list after the l_curr point - NOT at lhead.
* Note we can't use DLINK_ADD here as this inserts at the head of the given list.
*/
- ul_new->prev = ul_curr;
- ul_new->next = ul_curr->next;
- ul_curr->next = ul_new;
+ l_new->prev = l_curr;
+ l_new->next = l_curr->next;
+ l_curr->next = l_new;
/* And move after the link we added. */
- ul_curr = ul_new->next;
+ l_curr = l_new->next;
} else {
@@ -1201,18 +1005,159 @@ new: start=%.0f,size=%.0f\n", (double)ul_curr->start, (double)ul_curr->size,
*/
pstring msg;
- slprintf(msg, sizeof(msg)-1, "logic flaw in cases: ul_curr: start = %.0f, size = %.0f : \
-lock: start = %.0f, size = %.0f\n", (double)ul_curr->start, (double)ul_curr->size, (double)lock->start, (double)lock->size );
+ slprintf(msg, sizeof(msg)-1, "logic flaw in cases: l_curr: start = %.0f, size = %.0f : \
+lock: start = %.0f, size = %.0f\n", (double)l_curr->start, (double)l_curr->size, (double)lock->start, (double)lock->size );
smb_panic(msg);
}
- } /* end for ( ul_curr = ulhead; ul_curr;) */
+ } /* end for ( l_curr = lhead; l_curr;) */
} /* end for (i=0; i<num_locks && ul_head; i++) */
if (dbuf.dptr)
free(dbuf.dptr);
- return ulhead;
+ return lhead;
+}
+
+/****************************************************************************
+ POSIX function to acquire a lock. Returns True if the
+ lock could be granted, False if not.
+****************************************************************************/
+
+BOOL set_posix_lock(files_struct *fsp, SMB_BIG_UINT u_offset, SMB_BIG_UINT u_count, enum brl_type lock_type)
+{
+ SMB_OFF_T offset;
+ SMB_OFF_T count;
+ BOOL ret = True;
+ size_t entry_num = 0;
+ size_t lock_count;
+ TALLOC_CTX *l_ctx = NULL;
+ struct lock_list *llist = NULL;
+ struct lock_list *ll = NULL;
+ int posix_lock_type = map_posix_lock_type(fsp,lock_type);
+
+ DEBUG(5,("set_posix_lock: File %s, offset = %.0f, count = %.0f, type = %s\n",
+ fsp->fsp_name, (double)u_offset, (double)u_count, posix_lock_type_name(lock_type) ));
+
+ /*
+ * If the requested lock won't fit in the POSIX range, we will
+ * pretend it was successful.
+ */
+
+ if(!posix_lock_in_range(&offset, &count, u_offset, u_count))
+ return True;
+
+ /*
+ * Windows is very strange. It allows read locks to be overlayed
+ * (even over a write lock), but leaves the write lock in force until the first
+ * unlock. It also reference counts the locks. This means the following sequence :
+ *
+ * process1 process2
+ * ------------------------------------------------------------------------
+ * WRITE LOCK : start = 2, len = 10
+ * READ LOCK: start =0, len = 10 - FAIL
+ * READ LOCK : start = 0, len = 14
+ * READ LOCK: start =0, len = 10 - FAIL
+ * UNLOCK : start = 2, len = 10
+ * READ LOCK: start =0, len = 10 - OK
+ *
+ * Under POSIX, the same sequence in steps 1 and 2 would not be reference counted, but
+ * would leave a single read lock over the 0-14 region. In order to
+ * re-create Windows semantics mapped to POSIX locks, we create multiple TDB
+ * entries, one for each overlayed lock request. We are guarenteed by the brlock
+ * semantics that if a write lock is added, then it will be first in the array.
+ */
+
+ if ((l_ctx = talloc_init()) == NULL) {
+ DEBUG(0,("set_posix_lock: unable to init talloc context.\n"));
+ return True; /* Not a fatal error. */
+ }
+
+ if ((ll = (struct lock_list *)talloc(l_ctx, sizeof(struct lock_list))) == NULL) {
+ DEBUG(0,("set_posix_lock: unable to talloc unlock list.\n"));
+ talloc_destroy(l_ctx);
+ return True; /* Not a fatal error. */
+ }
+
+ /*
+ * Create the initial list entry containing the
+ * lock we want to add.
+ */
+
+ ZERO_STRUCTP(ll);
+ ll->start = offset;
+ ll->size = count;
+
+ DLIST_ADD(llist, ll);
+
+ /*
+ * The following call calculates if there are any
+ * overlapping locks held by this process on
+ * fd's open on the same file and splits this list
+ * into a list of lock ranges that do not overlap with existing
+ * POSIX locks.
+ */
+
+ llist = posix_lock_list(l_ctx, llist, fsp);
+
+ /*
+ * Now we have the list of ranges to lock it is safe to add the
+ * entry into the POSIX lock tdb. We take note of the entry we
+ * added here in case we have to remove it on POSIX lock fail.
+ */
+
+ if (!add_posix_lock_entry(fsp,offset,count,posix_lock_type,&entry_num)) {
+ DEBUG(0,("set_posix_lock: Unable to create posix lock entry !\n"));
+ talloc_destroy(l_ctx);
+ return False;
+ }
+
+ /*
+ * Add the POSIX locks on the list of ranges returned.
+ * As the lock is supposed to be added atomically, we need to
+ * back out all the locks if any one of these calls fail.
+ */
+
+ for (lock_count = 0, ll = llist; ll; ll = ll->next, lock_count++) {
+ offset = ll->start;
+ count = ll->size;
+
+ DEBUG(5,("set_posix_lock: Real lock: Type = %s: offset = %.0f, count = %.0f\n",
+ posix_lock_type_name(posix_lock_type), (double)offset, (double)count ));
+
+ if (!posix_fcntl_lock(fsp,SMB_F_SETLK,offset,count,posix_lock_type)) {
+ DEBUG(5,("set_posix_lock: Lock fail !: Type = %s: offset = %.0f, count = %.0f\n",
+ posix_lock_type_name(posix_lock_type), (double)offset, (double)count ));
+ ret = False;
+ break;
+ }
+ }
+
+ if (!ret) {
+
+ /*
+ * Back out all the POSIX locks we have on fail.
+ */
+
+ for (ll = llist; lock_count; ll = ll->next, lock_count--) {
+ offset = ll->start;
+ count = ll->size;
+
+ DEBUG(5,("set_posix_lock: Backing out locks: Type = %s: offset = %.0f, count = %.0f\n",
+ posix_lock_type_name(posix_lock_type), (double)offset, (double)count ));
+
+ posix_fcntl_lock(fsp,SMB_F_SETLK,offset,count,F_UNLCK);
+ }
+
+ /*
+ * Remove the tdb entry for this lock.
+ */
+
+ delete_posix_lock_entry_by_index(fsp,entry_num);
+ }
+
+ talloc_destroy(l_ctx);
+ return ret;
}
/****************************************************************************
@@ -1226,8 +1171,8 @@ BOOL release_posix_lock(files_struct *fsp, SMB_BIG_UINT u_offset, SMB_BIG_UINT u
SMB_OFF_T count;
BOOL ret = True;
TALLOC_CTX *ul_ctx = NULL;
- struct unlock_list *ulist = NULL;
- struct unlock_list *ul = NULL;
+ struct lock_list *ulist = NULL;
+ struct lock_list *ul = NULL;
struct posix_lock deleted_lock;
int num_overlapped_entries;
@@ -1245,6 +1190,8 @@ BOOL release_posix_lock(files_struct *fsp, SMB_BIG_UINT u_offset, SMB_BIG_UINT u
/*
* We treat this as one unlock request for POSIX accounting purposes even
* if it may later be split into multiple smaller POSIX unlock ranges.
+ * num_overlapped_entries is the number of existing locks that have any
+ * overlap with this unlock request.
*/
num_overlapped_entries = delete_posix_lock_entry(fsp, offset, count, &deleted_lock);
@@ -1256,7 +1203,8 @@ BOOL release_posix_lock(files_struct *fsp, SMB_BIG_UINT u_offset, SMB_BIG_UINT u
/*
* If num_overlapped_entries is > 0, and the lock_type we just deleted from the tdb was
* a POSIX write lock, then before doing the unlock we need to downgrade
- * the POSIX lock to a read lock.
+ * the POSIX lock to a read lock. This allows any overlapping read locks
+ * to be atomically maintained.
*/
if (num_overlapped_entries > 0 && deleted_lock.lock_type == F_WRLCK) {
@@ -1267,11 +1215,11 @@ BOOL release_posix_lock(files_struct *fsp, SMB_BIG_UINT u_offset, SMB_BIG_UINT u
}
if ((ul_ctx = talloc_init()) == NULL) {
- DEBUG(0,("release_posix_lock: unable to init talloc context.\n"));
+ DEBUG(0,("release_posix_lock: unable to init talloc context.\n"));
return True; /* Not a fatal error. */
}
- if ((ul = (struct unlock_list *)talloc(ul_ctx, sizeof(struct unlock_list))) == NULL) {
+ if ((ul = (struct lock_list *)talloc(ul_ctx, sizeof(struct lock_list))) == NULL) {
DEBUG(0,("release_posix_lock: unable to talloc unlock list.\n"));
talloc_destroy(ul_ctx);
return True; /* Not a fatal error. */
@@ -1297,7 +1245,7 @@ BOOL release_posix_lock(files_struct *fsp, SMB_BIG_UINT u_offset, SMB_BIG_UINT u
* unlocks are performed.
*/
- ulist = posix_unlock_list(ul_ctx, ulist, fsp);
+ ulist = posix_lock_list(ul_ctx, ulist, fsp);
/*
* Release the POSIX locks on the list of ranges returned.
@@ -1307,16 +1255,6 @@ BOOL release_posix_lock(files_struct *fsp, SMB_BIG_UINT u_offset, SMB_BIG_UINT u
offset = ulist->start;
count = ulist->size;
- if(u_count == 0) {
-
- /*
- * This lock must overlap with an existing lock.
- * Don't do any POSIX call.
- */
-
- continue;
- }
-
DEBUG(5,("release_posix_lock: Real unlock: offset = %.0f, count = %.0f\n",
(double)offset, (double)count ));