Mercurial > dovecot > original-hg > dovecot-1.2
view src/lib-index/mail-index-lock.c @ 2936:57f975a9b63b HEAD
Renamed mail_index_refresh() to mail_index_reopen_if_needed(). Added public
mail_index_refresh() which makes sure index is fully refreshed at the time.
Added mbox code to call it after mbox is locked to avoid using old mbox offsets.
author | Timo Sirainen <tss@iki.fi> |
---|---|
date | Sun, 05 Dec 2004 03:45:53 +0200 |
parents | 173a0296e81d |
children | 89ab5c72430d |
line wrap: on
line source
/* Copyright (C) 2003-2004 Timo Sirainen */ /* Locking should never fail or timeout. Exclusive locks must be kept as short time as possible. Shared locks can be long living, so if we can't get exclusive lock directly within 2 seconds, we'll replace the index file with a copy of it. That means the shared lock holders can keep using the old file while we're modifying the new file. lock_id is used to figure out if acquired lock is still valid. When index file is reopened, the lock_id can become invalid. It doesn't matter however, as no-one's going to modify the old file anymore. lock_id also tells if we're referring to shared or exclusive lock. This allows us to drop back to shared locking once all exclusive locks are dropped. Shared locks have even numbers, exclusive locks have odd numbers. The number is increased by two every time the lock is dropped or index file is reopened. */ #include "lib.h" #include "buffer.h" #include "mmap-util.h" #include "write-full.h" #include "mail-index-private.h" #include <stdio.h> #include <sys/stat.h> #ifdef HAVE_FLOCK # include <sys/file.h> #endif #define MAIL_INDEX_LOCK_WAIT_TIME 120 int mail_index_lock_fd(struct mail_index *index, int fd, int lock_type, unsigned int timeout_secs) { if (timeout_secs != 0) alarm(MAIL_INDEX_LOCK_WAIT_TIME); switch (index->lock_method) { case MAIL_INDEX_LOCK_FCNTL: { #ifndef HAVE_FCNTL i_fatal("fcntl() locks not supported"); #else struct flock fl; fl.l_type = lock_type; fl.l_whence = SEEK_SET; fl.l_start = 0; fl.l_len = 0; if (fcntl(fd, timeout_secs ? F_SETLKW : F_SETLK, &fl) < 0) { if (timeout_secs == 0 && (errno == EACCES || errno == EAGAIN)) { /* locked by another process */ return 0; } if (errno == EINTR) { /* most likely alarm hit, meaning we timeouted. even if not, we probably want to be killed so stop blocking. */ errno = EAGAIN; if (timeout_secs != 0) alarm(0); return 0; } } if (timeout_secs != 0) alarm(0); return 1; #endif } case MAIL_INDEX_LOCK_FLOCK: { #ifndef HAVE_FLOCK i_fatal("flock() locks not supported"); #else int operation = timeout_secs != 0 ? 0 : LOCK_NB; switch (lock_type) { case F_RDLCK: operation |= LOCK_SH; break; case F_WRLCK: operation |= LOCK_EX; break; case F_UNLCK: operation |= LOCK_UN; break; } if (flock(fd, operation) < 0) { if (errno == EWOULDBLOCK || errno == EINTR) { /* a) locked by another process, b) timeouted */ if (timeout_secs != 0) alarm(0); return 0; } } if (timeout_secs != 0) alarm(0); return 1; #endif } case MAIL_INDEX_LOCK_DOTLOCK: /* we shouldn't get here */ break; } i_unreached(); } static int mail_index_lock(struct mail_index *index, int lock_type, unsigned int timeout_secs, int update_index, unsigned int *lock_id_r) { int ret, ret2; i_assert(lock_type == F_RDLCK || lock_type == F_WRLCK); if (lock_type == F_RDLCK && index->lock_type != F_UNLCK) { index->shared_lock_count++; *lock_id_r = index->lock_id; ret = 1; } else if (lock_type == F_WRLCK && index->lock_type == F_WRLCK) { index->excl_lock_count++; *lock_id_r = index->lock_id + 1; ret = 1; } else { ret = 0; } if (update_index && index->excl_lock_count == 0) { i_assert(index->lock_type != F_WRLCK); if ((ret2 = mail_index_reopen_if_needed(index)) < 0) return -1; if (ret > 0 && ret2 == 0) { i_assert(lock_type == F_RDLCK); i_assert(index->lock_type == F_RDLCK); return 1; } ret = 0; } if (ret > 0) return 1; if (index->lock_method == MAIL_INDEX_LOCK_DOTLOCK) { /* FIXME: exclusive locking will rewrite the index file every time. shouldn't really be needed.. reading doesn't require locks then, though */ if (lock_type == F_WRLCK) return 0; if (update_index && index->lock_type == F_UNLCK) { if (mail_index_reopen_if_needed(index) < 0) return -1; } index->shared_lock_count++; index->lock_type = F_RDLCK; *lock_id_r = index->lock_id; return 1; } if (lock_type == F_RDLCK || !index->log_locked) { ret = mail_index_lock_fd(index, index->fd, lock_type, timeout_secs); } else { /* this is kind of kludgy. we wish to avoid deadlocks while trying to lock transaction log, but it can happen if our process is holding transaction log lock and waiting for index write lock, while the other process is holding index read lock and waiting for transaction log lock. we don't have a problem with grabbing read index lock because the only way for it to block is if it's write-locked, which isn't allowed unless transaction log is also locked. so, the workaround for this problem is that we simply try locking once. if it doesn't work, just rewrite the file. hopefully there won't be any other deadlocking issues. :) */ ret = mail_index_lock_fd(index, index->fd, lock_type, 0); } if (ret <= 0) { if (ret == 0) return 0; mail_index_set_syscall_error(index, "mail_index_lock_fd()"); return -1; } if (index->lock_type == F_UNLCK) index->lock_id += 2; index->lock_type = lock_type; if (lock_type == F_RDLCK) { index->shared_lock_count++; *lock_id_r = index->lock_id; } else { index->excl_lock_count++; *lock_id_r = index->lock_id + 1; } return 1; } int mail_index_lock_shared(struct mail_index *index, int update_index, unsigned int *lock_id_r) { int ret; ret = mail_index_lock(index, F_RDLCK, MAIL_INDEX_LOCK_SECS, update_index, lock_id_r); if (ret > 0) return 0; if (ret < 0) return -1; mail_index_set_error(index, "Timeout while waiting for release of " "shared fcntl() lock for index file %s", index->filepath); index->index_lock_timeout = TRUE; return -1; } static int mail_index_copy(struct mail_index *index) { const char *path; int ret, fd; fd = mail_index_create_tmp_file(index, &path); if (fd == -1) return -1; /* write base header */ ret = write_full(fd, &index->map->hdr, index->map->hdr.base_header_size); if (ret == 0) { /* write extended headers */ const struct mail_index_header *hdr; const void *hdr_ext; hdr = index->map->hdr_base; hdr_ext = CONST_PTR_OFFSET(hdr, hdr->base_header_size); ret = write_full(fd, hdr_ext, hdr->header_size - hdr->base_header_size); } if (ret < 0 || write_full(fd, index->map->records, index->map->records_count * index->map->hdr.record_size) < 0) { mail_index_file_set_syscall_error(index, path, "write_full()"); (void)close(fd); (void)unlink(path); fd = -1; } else { i_assert(index->copy_lock_path == NULL); index->copy_lock_path = i_strdup(path); } return fd; } static int mail_index_lock_exclusive_copy(struct mail_index *index) { int fd, old_lock_type; i_assert(index->log_locked); if (index->copy_lock_path != NULL) { index->excl_lock_count++; return 0; } i_assert(index->excl_lock_count == 0); /* copy the index to index.tmp and use it */ fd = mail_index_copy(index); if (fd == -1) return -1; old_lock_type = index->lock_type; index->lock_type = F_WRLCK; index->excl_lock_count++; if (mail_index_reopen(index, fd) < 0) { i_assert(index->excl_lock_count == 1); (void)close(fd); if (unlink(index->copy_lock_path) < 0) { mail_index_file_set_syscall_error(index, index->copy_lock_path, "unlink()"); } i_free(index->copy_lock_path); index->copy_lock_path = NULL; index->lock_type = old_lock_type; index->excl_lock_count = 0; return -1; } return 0; } int mail_index_lock_exclusive(struct mail_index *index, unsigned int *lock_id_r) { int ret; /* exclusive transaction log lock protects exclusive locking for the main index file */ i_assert(index->log_locked); /* if header size is smaller than what we have, we'll have to recreate the index to grow it. so don't even try regular locking. */ if (index->map->hdr.base_header_size >= sizeof(*index->hdr) || index->excl_lock_count > 0) { /* wait two seconds for exclusive lock */ ret = mail_index_lock(index, F_WRLCK, 2, TRUE, lock_id_r); if (ret > 0) return 0; if (ret < 0) return -1; } if (mail_index_lock_exclusive_copy(index) < 0) return -1; *lock_id_r = index->lock_id + 1; return 0; } static int mail_index_copy_lock_finish(struct mail_index *index) { if (fsync(index->fd) < 0) { mail_index_file_set_syscall_error(index, index->copy_lock_path, "fsync()"); return -1; } if (rename(index->copy_lock_path, index->filepath) < 0) { mail_index_set_error(index, "rename(%s, %s) failed: %m", index->copy_lock_path, index->filepath); return -1; } i_free(index->copy_lock_path); index->copy_lock_path = NULL; return 0; } static void mail_index_excl_unlock_finish(struct mail_index *index) { int fd; if (index->map != NULL && index->map->write_to_disk) { i_assert(index->log_locked); index->map->write_to_disk = FALSE; if (index->copy_lock_path != NULL) { /* new mapping replaces the old */ (void)unlink(index->copy_lock_path); i_free(index->copy_lock_path); index->copy_lock_path = NULL; } fd = mail_index_copy(index); if (fd == -1) mail_index_set_inconsistent(index); else (void)close(fd); } if (index->shared_lock_count > 0 && index->lock_method != MAIL_INDEX_LOCK_DOTLOCK) { /* leave ourself shared locked. */ if (mail_index_lock_fd(index, index->fd, F_RDLCK, 0) <= 0) { mail_index_file_set_syscall_error(index, index->copy_lock_path, "mail_index_lock_fd()"); } i_assert(index->lock_type == F_WRLCK); index->lock_type = F_RDLCK; } if (index->copy_lock_path != NULL) { i_assert(index->log_locked); if (mail_index_copy_lock_finish(index) < 0) mail_index_set_inconsistent(index); } } void mail_index_unlock(struct mail_index *index, unsigned int lock_id) { if ((lock_id & 1) == 0) { /* shared lock */ if (!mail_index_is_locked(index, lock_id)) { /* unlocking some older generation of the index file. we've already closed the file so just ignore this. */ return; } i_assert(index->shared_lock_count > 0); index->shared_lock_count--; } else { /* exclusive lock */ i_assert(lock_id == index->lock_id + 1); i_assert(index->excl_lock_count > 0); if (--index->excl_lock_count == 0) mail_index_excl_unlock_finish(index); } if (index->shared_lock_count == 0 && index->excl_lock_count == 0) { index->lock_id += 2; index->lock_type = F_UNLCK; if (index->lock_method != MAIL_INDEX_LOCK_DOTLOCK) { if (mail_index_lock_fd(index, index->fd, F_UNLCK, 0) < 0) { mail_index_set_syscall_error(index, "mail_index_lock_fd()"); } } } } int mail_index_is_locked(struct mail_index *index, unsigned int lock_id) { if ((index->lock_id ^ lock_id) <= 1) { i_assert(index->lock_type != F_UNLCK); return TRUE; } return FALSE; }