/* * Simulate the Posix AIO using mmap/fork * * Copyright (C) Volker Lendecke 2008 * Copyright (C) Jeremy Allison 2010 * * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation; either version 2 of the License, or * (at your option) any later version. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, write to the Free Software * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA. */ #include "includes.h" #include "system/filesys.h" #include "system/shmem.h" #include "smbd/smbd.h" #include "smbd/globals.h" #include "lib/async_req/async_sock.h" #include "lib/util/tevent_unix.h" #undef recvmsg #ifndef MAP_FILE #define MAP_FILE 0 #endif struct aio_fork_config { bool erratic_testing_mode; }; struct mmap_area { size_t size; volatile void *ptr; }; static int mmap_area_destructor(struct mmap_area *area) { munmap((void *)area->ptr, area->size); return 0; } static struct mmap_area *mmap_area_init(TALLOC_CTX *mem_ctx, size_t size) { struct mmap_area *result; int fd; result = talloc(mem_ctx, struct mmap_area); if (result == NULL) { DEBUG(0, ("talloc failed\n")); goto fail; } fd = open("/dev/zero", O_RDWR); if (fd == -1) { DEBUG(3, ("open(\"/dev/zero\") failed: %s\n", strerror(errno))); goto fail; } result->ptr = mmap(NULL, size, PROT_READ|PROT_WRITE, MAP_SHARED|MAP_FILE, fd, 0); if (result->ptr == MAP_FAILED) { DEBUG(1, ("mmap failed: %s\n", strerror(errno))); goto fail; } close(fd); result->size = size; talloc_set_destructor(result, mmap_area_destructor); return result; fail: TALLOC_FREE(result); return NULL; } enum cmd_type { READ_CMD, WRITE_CMD, FSYNC_CMD }; static const char *cmd_type_str(enum cmd_type cmd) { const char *result; switch (cmd) { case READ_CMD: result = "READ"; break; case WRITE_CMD: result = "WRITE"; break; case FSYNC_CMD: result = "FSYNC"; break; default: result = ""; break; } return result; } struct rw_cmd { size_t n; off_t offset; enum cmd_type cmd; bool erratic_testing_mode; }; struct rw_ret { ssize_t size; int ret_errno; }; struct aio_child_list; struct aio_child { struct aio_child *prev, *next; struct aio_child_list *list; pid_t pid; int sockfd; struct mmap_area *map; bool dont_delete; /* Marked as in use since last cleanup */ bool busy; }; struct aio_child_list { struct aio_child *children; struct tevent_timer *cleanup_event; }; static void free_aio_children(void **p) { TALLOC_FREE(*p); } static ssize_t read_fd(int fd, void *ptr, size_t nbytes, int *recvfd) { struct msghdr msg; struct iovec iov[1]; ssize_t n; #ifndef HAVE_MSGHDR_MSG_CONTROL int newfd; #endif #ifdef HAVE_MSGHDR_MSG_CONTROL union { struct cmsghdr cm; char control[CMSG_SPACE(sizeof(int))]; } control_un; struct cmsghdr *cmptr; msg.msg_control = control_un.control; msg.msg_controllen = sizeof(control_un.control); #else #if HAVE_MSGHDR_MSG_ACCTRIGHTS msg.msg_accrights = (caddr_t) &newfd; msg.msg_accrightslen = sizeof(int); #else #error Can not pass file descriptors #endif #endif msg.msg_name = NULL; msg.msg_namelen = 0; msg.msg_flags = 0; iov[0].iov_base = (void *)ptr; iov[0].iov_len = nbytes; msg.msg_iov = iov; msg.msg_iovlen = 1; if ( (n = recvmsg(fd, &msg, 0)) <= 0) { return(n); } #ifdef HAVE_MSGHDR_MSG_CONTROL if ((cmptr = CMSG_FIRSTHDR(&msg)) != NULL && cmptr->cmsg_len == CMSG_LEN(sizeof(int))) { if (cmptr->cmsg_level != SOL_SOCKET) { DEBUG(10, ("control level != SOL_SOCKET")); errno = EINVAL; return -1; } if (cmptr->cmsg_type != SCM_RIGHTS) { DEBUG(10, ("control type != SCM_RIGHTS")); errno = EINVAL; return -1; } memcpy(recvfd, CMSG_DATA(cmptr), sizeof(*recvfd)); } else { *recvfd = -1; /* descriptor was not passed */ } #else if (msg.msg_accrightslen == sizeof(int)) { *recvfd = newfd; } else { *recvfd = -1; /* descriptor was not passed */ } #endif return(n); } static ssize_t write_fd(int fd, void *ptr, size_t nbytes, int sendfd) { struct msghdr msg; struct iovec iov[1]; #ifdef HAVE_MSGHDR_MSG_CONTROL union { struct cmsghdr cm; char control[CMSG_SPACE(sizeof(int))]; } control_un; struct cmsghdr *cmptr; ZERO_STRUCT(msg); ZERO_STRUCT(control_un); msg.msg_control = control_un.control; msg.msg_controllen = sizeof(control_un.control); cmptr = CMSG_FIRSTHDR(&msg); cmptr->cmsg_len = CMSG_LEN(sizeof(int)); cmptr->cmsg_level = SOL_SOCKET; cmptr->cmsg_type = SCM_RIGHTS; memcpy(CMSG_DATA(cmptr), &sendfd, sizeof(sendfd)); #else ZERO_STRUCT(msg); msg.msg_accrights = (caddr_t) &sendfd; msg.msg_accrightslen = sizeof(int); #endif msg.msg_name = NULL; msg.msg_namelen = 0; ZERO_STRUCT(iov); iov[0].iov_base = (void *)ptr; iov[0].iov_len = nbytes; msg.msg_iov = iov; msg.msg_iovlen = 1; return (sendmsg(fd, &msg, 0)); } static void aio_child_cleanup(struct tevent_context *event_ctx, struct tevent_timer *te, struct timeval now, void *private_data) { struct aio_child_list *list = talloc_get_type_abort( private_data, struct aio_child_list); struct aio_child *child, *next; TALLOC_FREE(list->cleanup_event); for (child = list->children; child != NULL; child = next) { next = child->next; if (child->busy) { DEBUG(10, ("child %d currently active\n", (int)child->pid)); continue; } if (child->dont_delete) { DEBUG(10, ("Child %d was active since last cleanup\n", (int)child->pid)); child->dont_delete = false; continue; } DEBUG(10, ("Child %d idle for more than 30 seconds, " "deleting\n", (int)child->pid)); TALLOC_FREE(child); child = next; } if (list->children != NULL) { /* * Re-schedule the next cleanup round */ list->cleanup_event = tevent_add_timer(server_event_context(), list, timeval_add(&now, 30, 0), aio_child_cleanup, list); } } static struct aio_child_list *init_aio_children(struct vfs_handle_struct *handle) { struct aio_child_list *data = NULL; if (SMB_VFS_HANDLE_TEST_DATA(handle)) { SMB_VFS_HANDLE_GET_DATA(handle, data, struct aio_child_list, return NULL); } if (data == NULL) { data = talloc_zero(NULL, struct aio_child_list); if (data == NULL) { return NULL; } } /* * Regardless of whether the child_list had been around or not, make * sure that we have a cleanup timed event. This timed event will * delete itself when it finds that no children are around anymore. */ if (data->cleanup_event == NULL) { data->cleanup_event = tevent_add_timer(server_event_context(), data, timeval_current_ofs(30, 0), aio_child_cleanup, data); if (data->cleanup_event == NULL) { TALLOC_FREE(data); return NULL; } } if (!SMB_VFS_HANDLE_TEST_DATA(handle)) { SMB_VFS_HANDLE_SET_DATA(handle, data, free_aio_children, struct aio_child_list, return False); } return data; } static void aio_child_loop(int sockfd, struct mmap_area *map) { while (true) { int fd = -1; ssize_t ret; struct rw_cmd cmd_struct; struct rw_ret ret_struct; ret = read_fd(sockfd, &cmd_struct, sizeof(cmd_struct), &fd); if (ret != sizeof(cmd_struct)) { DEBUG(10, ("read_fd returned %d: %s\n", (int)ret, strerror(errno))); exit(1); } DEBUG(10, ("aio_child_loop: %s %d bytes at %d from fd %d\n", cmd_type_str(cmd_struct.cmd), (int)cmd_struct.n, (int)cmd_struct.offset, fd)); if (cmd_struct.erratic_testing_mode) { /* * For developer testing, we want erratic behaviour for * async I/O times */ uint8_t randval; unsigned msecs; /* * use generate_random_buffer, we just forked from a * common parent state */ generate_random_buffer(&randval, sizeof(randval)); msecs = randval + 20; DEBUG(10, ("delaying for %u msecs\n", msecs)); smb_msleep(msecs); } ZERO_STRUCT(ret_struct); switch (cmd_struct.cmd) { case READ_CMD: ret_struct.size = sys_pread( fd, (void *)map->ptr, cmd_struct.n, cmd_struct.offset); #if 0 /* This breaks "make test" when run with aio_fork module. */ #ifdef DEVELOPER ret_struct.size = MAX(1, ret_struct.size * 0.9); #endif #endif break; case WRITE_CMD: ret_struct.size = sys_pwrite( fd, (void *)map->ptr, cmd_struct.n, cmd_struct.offset); break; case FSYNC_CMD: ret_struct.size = fsync(fd); break; default: ret_struct.size = -1; errno = EINVAL; } DEBUG(10, ("aio_child_loop: syscall returned %d\n", (int)ret_struct.size)); if (ret_struct.size == -1) { ret_struct.ret_errno = errno; } /* * Close the fd before telling our parent we're done. The * parent might close and re-open the file very quickly, and * with system-level share modes (GPFS) we would get an * unjustified SHARING_VIOLATION. */ close(fd); ret = write_data(sockfd, (char *)&ret_struct, sizeof(ret_struct)); if (ret != sizeof(ret_struct)) { DEBUG(10, ("could not write ret_struct: %s\n", strerror(errno))); exit(2); } } } static int aio_child_destructor(struct aio_child *child) { char c=0; SMB_ASSERT(!child->busy); DEBUG(10, ("aio_child_destructor: removing child %d on fd %d\n", child->pid, child->sockfd)); /* * closing the sockfd makes the child not return from recvmsg() on RHEL * 5.5 so instead force the child to exit by writing bad data to it */ write(child->sockfd, &c, sizeof(c)); close(child->sockfd); DLIST_REMOVE(child->list->children, child); return 0; } /* * We have to close all fd's in open files, we might incorrectly hold a system * level share mode on a file. */ static struct files_struct *close_fsp_fd(struct files_struct *fsp, void *private_data) { if ((fsp->fh != NULL) && (fsp->fh->fd != -1)) { close(fsp->fh->fd); fsp->fh->fd = -1; } return NULL; } static int create_aio_child(struct smbd_server_connection *sconn, struct aio_child_list *children, size_t map_size, struct aio_child **presult) { struct aio_child *result; int fdpair[2]; int ret; fdpair[0] = fdpair[1] = -1; result = talloc_zero(children, struct aio_child); if (result == NULL) { return ENOMEM; } if (socketpair(AF_UNIX, SOCK_STREAM, 0, fdpair) == -1) { ret = errno; DEBUG(10, ("socketpair() failed: %s\n", strerror(errno))); goto fail; } DEBUG(10, ("fdpair = %d/%d\n", fdpair[0], fdpair[1])); result->map = mmap_area_init(result, map_size); if (result->map == NULL) { ret = errno; DEBUG(0, ("Could not create mmap area\n")); goto fail; } result->pid = fork(); if (result->pid == -1) { ret = errno; DEBUG(0, ("fork failed: %s\n", strerror(errno))); goto fail; } if (result->pid == 0) { close(fdpair[0]); result->sockfd = fdpair[1]; files_forall(sconn, close_fsp_fd, NULL); aio_child_loop(result->sockfd, result->map); } DEBUG(10, ("Child %d created with sockfd %d\n", result->pid, fdpair[0])); result->sockfd = fdpair[0]; close(fdpair[1]); result->list = children; DLIST_ADD(children->children, result); talloc_set_destructor(result, aio_child_destructor); *presult = result; return 0; fail: if (fdpair[0] != -1) close(fdpair[0]); if (fdpair[1] != -1) close(fdpair[1]); TALLOC_FREE(result); return ret; } static int get_idle_child(struct vfs_handle_struct *handle, struct aio_child **pchild) { struct aio_child_list *children; struct aio_child *child; children = init_aio_children(handle); if (children == NULL) { return ENOMEM; } for (child = children->children; child != NULL; child = child->next) { if (!child->busy) { break; } } if (child == NULL) { int ret; DEBUG(10, ("no idle child found, creating new one\n")); ret = create_aio_child(handle->conn->sconn, children, 128*1024, &child); if (ret != 0) { DEBUG(10, ("create_aio_child failed: %s\n", strerror(errno))); return ret; } } child->dont_delete = true; child->busy = true; *pchild = child; return 0; } struct aio_fork_pread_state { struct aio_child *child; ssize_t ret; int err; }; static void aio_fork_pread_done(struct tevent_req *subreq); static struct tevent_req *aio_fork_pread_send(struct vfs_handle_struct *handle, TALLOC_CTX *mem_ctx, struct tevent_context *ev, struct files_struct *fsp, void *data, size_t n, off_t offset) { struct tevent_req *req, *subreq; struct aio_fork_pread_state *state; struct rw_cmd cmd; ssize_t written; int err; struct aio_fork_config *config; SMB_VFS_HANDLE_GET_DATA(handle, config, struct aio_fork_config, return NULL); req = tevent_req_create(mem_ctx, &state, struct aio_fork_pread_state); if (req == NULL) { return NULL; } if (n > 128*1024) { /* TODO: support variable buffers */ tevent_req_error(req, EINVAL); return tevent_req_post(req, ev); } err = get_idle_child(handle, &state->child); if (err != 0) { tevent_req_error(req, err); return tevent_req_post(req, ev); } ZERO_STRUCT(cmd); cmd.n = n; cmd.offset = offset; cmd.cmd = READ_CMD; cmd.erratic_testing_mode = config->erratic_testing_mode; DEBUG(10, ("sending fd %d to child %d\n", fsp->fh->fd, (int)state->child->pid)); /* * Not making this async. We're writing into an empty unix * domain socket. This should never block. */ written = write_fd(state->child->sockfd, &cmd, sizeof(cmd), fsp->fh->fd); if (written == -1) { err = errno; TALLOC_FREE(state->child); DEBUG(10, ("write_fd failed: %s\n", strerror(err))); tevent_req_error(req, err); return tevent_req_post(req, ev); } subreq = read_packet_send(state, ev, state->child->sockfd, sizeof(struct rw_ret), NULL, NULL); if (tevent_req_nomem(subreq, req)) { TALLOC_FREE(state->child); /* we sent sth down */ return tevent_req_post(req, ev); } tevent_req_set_callback(subreq, aio_fork_pread_done, req); return req; } static void aio_fork_pread_done(struct tevent_req *subreq) { struct tevent_req *req = tevent_req_callback_data( subreq, struct tevent_req); struct aio_fork_pread_state *state = tevent_req_data( req, struct aio_fork_pread_state); ssize_t nread; uint8_t *buf; int err; struct rw_ret *retbuf; nread = read_packet_recv(subreq, talloc_tos(), &buf, &err); TALLOC_FREE(subreq); if (nread == -1) { TALLOC_FREE(state->child); tevent_req_error(req, err); return; } state->child->busy = false; retbuf = (struct rw_ret *)buf; state->ret = retbuf->size; state->err = retbuf->ret_errno; tevent_req_done(req); } static ssize_t aio_fork_pread_recv(struct tevent_req *req, int *err) { struct aio_fork_pread_state *state = tevent_req_data( req, struct aio_fork_pread_state); if (tevent_req_is_unix_error(req, err)) { return -1; } if (state->ret == -1) { *err = state->err; } return state->ret; } struct aio_fork_pwrite_state { struct aio_child *child; ssize_t ret; int err; }; static void aio_fork_pwrite_done(struct tevent_req *subreq); static struct tevent_req *aio_fork_pwrite_send( struct vfs_handle_struct *handle, TALLOC_CTX *mem_ctx, struct tevent_context *ev, struct files_struct *fsp, const void *data, size_t n, off_t offset) { struct tevent_req *req, *subreq; struct aio_fork_pwrite_state *state; struct rw_cmd cmd; ssize_t written; int err; struct aio_fork_config *config; SMB_VFS_HANDLE_GET_DATA(handle, config, struct aio_fork_config, return NULL); req = tevent_req_create(mem_ctx, &state, struct aio_fork_pwrite_state); if (req == NULL) { return NULL; } if (n > 128*1024) { /* TODO: support variable buffers */ tevent_req_error(req, EINVAL); return tevent_req_post(req, ev); } err = get_idle_child(handle, &state->child); if (err != 0) { tevent_req_error(req, err); return tevent_req_post(req, ev); } ZERO_STRUCT(cmd); cmd.n = n; cmd.offset = offset; cmd.cmd = WRITE_CMD; cmd.erratic_testing_mode = config->erratic_testing_mode; DEBUG(10, ("sending fd %d to child %d\n", fsp->fh->fd, (int)state->child->pid)); /* * Not making this async. We're writing into an empty unix * domain socket. This should never block. */ written = write_fd(state->child->sockfd, &cmd, sizeof(cmd), fsp->fh->fd); if (written == -1) { err = errno; TALLOC_FREE(state->child); DEBUG(10, ("write_fd failed: %s\n", strerror(err))); tevent_req_error(req, err); return tevent_req_post(req, ev); } subreq = read_packet_send(state, ev, state->child->sockfd, sizeof(struct rw_ret), NULL, NULL); if (tevent_req_nomem(subreq, req)) { TALLOC_FREE(state->child); /* we sent sth down */ return tevent_req_post(req, ev); } tevent_req_set_callback(subreq, aio_fork_pwrite_done, req); return req; } static void aio_fork_pwrite_done(struct tevent_req *subreq) { struct tevent_req *req = tevent_req_callback_data( subreq, struct tevent_req); struct aio_fork_pwrite_state *state = tevent_req_data( req, struct aio_fork_pwrite_state); ssize_t nread; uint8_t *buf; int err; struct rw_ret *retbuf; nread = read_packet_recv(subreq, talloc_tos(), &buf, &err); TALLOC_FREE(subreq); if (nread == -1) { TALLOC_FREE(state->child); tevent_req_error(req, err); return; } state->child->busy = false; retbuf = (struct rw_ret *)buf; state->ret = retbuf->size; state->err = retbuf->ret_errno; tevent_req_done(req); } static ssize_t aio_fork_pwrite_recv(struct tevent_req *req, int *err) { struct aio_fork_pwrite_state *state = tevent_req_data( req, struct aio_fork_pwrite_state); if (tevent_req_is_unix_error(req, err)) { return -1; } if (state->ret == -1) { *err = state->err; } return state->ret; } struct aio_fork_fsync_state { struct aio_child *child; ssize_t ret; int err; }; static void aio_fork_fsync_done(struct tevent_req *subreq); static struct tevent_req *aio_fork_fsync_send( struct vfs_handle_struct *handle, TALLOC_CTX *mem_ctx, struct tevent_context *ev, struct files_struct *fsp) { struct tevent_req *req, *subreq; struct aio_fork_fsync_state *state; struct rw_cmd cmd; ssize_t written; int err; struct aio_fork_config *config; SMB_VFS_HANDLE_GET_DATA(handle, config, struct aio_fork_config, return NULL); req = tevent_req_create(mem_ctx, &state, struct aio_fork_fsync_state); if (req == NULL) { return NULL; } err = get_idle_child(handle, &state->child); if (err != 0) { tevent_req_error(req, err); return tevent_req_post(req, ev); } ZERO_STRUCT(cmd); cmd.cmd = FSYNC_CMD; cmd.erratic_testing_mode = config->erratic_testing_mode; DEBUG(10, ("sending fd %d to child %d\n", fsp->fh->fd, (int)state->child->pid)); /* * Not making this async. We're writing into an empty unix * domain socket. This should never block. */ written = write_fd(state->child->sockfd, &cmd, sizeof(cmd), fsp->fh->fd); if (written == -1) { err = errno; TALLOC_FREE(state->child); DEBUG(10, ("write_fd failed: %s\n", strerror(err))); tevent_req_error(req, err); return tevent_req_post(req, ev); } subreq = read_packet_send(state, ev, state->child->sockfd, sizeof(struct rw_ret), NULL, NULL); if (tevent_req_nomem(subreq, req)) { TALLOC_FREE(state->child); /* we sent sth down */ return tevent_req_post(req, ev); } tevent_req_set_callback(subreq, aio_fork_fsync_done, req); return req; } static void aio_fork_fsync_done(struct tevent_req *subreq) { struct tevent_req *req = tevent_req_callback_data( subreq, struct tevent_req); struct aio_fork_fsync_state *state = tevent_req_data( req, struct aio_fork_fsync_state); ssize_t nread; uint8_t *buf; int err; struct rw_ret *retbuf; nread = read_packet_recv(subreq, talloc_tos(), &buf, &err); TALLOC_FREE(subreq); if (nread == -1) { TALLOC_FREE(state->child); tevent_req_error(req, err); return; } state->child->busy = false; retbuf = (struct rw_ret *)buf; state->ret = retbuf->size; state->err = retbuf->ret_errno; tevent_req_done(req); } static int aio_fork_fsync_recv(struct tevent_req *req, int *err) { struct aio_fork_fsync_state *state = tevent_req_data( req, struct aio_fork_fsync_state); if (tevent_req_is_unix_error(req, err)) { return -1; } if (state->ret == -1) { *err = state->err; } return state->ret; } static int aio_fork_connect(vfs_handle_struct *handle, const char *service, const char *user) { int ret; struct aio_fork_config *config; ret = SMB_VFS_NEXT_CONNECT(handle, service, user); if (ret < 0) { return ret; } config = talloc_zero(handle->conn, struct aio_fork_config); if (!config) { SMB_VFS_NEXT_DISCONNECT(handle); DEBUG(0, ("talloc_zero() failed\n")); return -1; } config->erratic_testing_mode = lp_parm_bool(SNUM(handle->conn), "vfs_aio_fork", "erratic_testing_mode", false); SMB_VFS_HANDLE_SET_DATA(handle, config, NULL, struct aio_fork_config, return -1); /********************************************************************* * How many threads to initialize ? * 100 per process seems insane as a default until you realize that * (a) Threads terminate after 1 second when idle. * (b) Throttling is done in SMB2 via the crediting algorithm. * (c) SMB1 clients are limited to max_mux (50) outstanding * requests and Windows clients don't use this anyway. * Essentially we want this to be unlimited unless smb.conf * says different. *********************************************************************/ aio_pending_size = 100; return 0; } static struct vfs_fn_pointers vfs_aio_fork_fns = { .connect_fn = aio_fork_connect, .pread_send_fn = aio_fork_pread_send, .pread_recv_fn = aio_fork_pread_recv, .pwrite_send_fn = aio_fork_pwrite_send, .pwrite_recv_fn = aio_fork_pwrite_recv, .fsync_send_fn = aio_fork_fsync_send, .fsync_recv_fn = aio_fork_fsync_recv, }; NTSTATUS vfs_aio_fork_init(void); NTSTATUS vfs_aio_fork_init(void) { return smb_register_vfs(SMB_VFS_INTERFACE_VERSION, "aio_fork", &vfs_aio_fork_fns); }