2012-06-06 10:45:07 +01:00
|
|
|
#include "serve.h"
|
2012-06-06 11:27:52 +01:00
|
|
|
#include "client.h"
|
2012-05-17 20:14:22 +01:00
|
|
|
#include "nbdtypes.h"
|
|
|
|
#include "ioutil.h"
|
|
|
|
#include "util.h"
|
2012-05-18 13:24:35 +01:00
|
|
|
#include "bitset.h"
|
2012-05-23 00:42:14 +01:00
|
|
|
#include "control.h"
|
2012-06-06 12:41:03 +01:00
|
|
|
#include "self_pipe.h"
|
2012-05-17 20:14:22 +01:00
|
|
|
|
|
|
|
#include <sys/types.h>
|
|
|
|
#include <sys/stat.h>
|
|
|
|
#include <sys/mman.h>
|
2012-05-18 18:44:34 +01:00
|
|
|
#include <sys/un.h>
|
2012-05-17 20:14:22 +01:00
|
|
|
#include <fcntl.h>
|
|
|
|
|
|
|
|
#include <string.h>
|
|
|
|
#include <stdlib.h>
|
|
|
|
#include <errno.h>
|
|
|
|
|
2012-05-31 11:33:31 +01:00
|
|
|
#include <sys/socket.h>
|
|
|
|
#include <netinet/tcp.h>
|
|
|
|
|
2012-06-06 10:45:07 +01:00
|
|
|
static inline void* sockaddr_address_data(struct sockaddr* sockaddr)
|
|
|
|
{
|
2012-06-07 11:44:19 +01:00
|
|
|
NULLCHECK( sockaddr );
|
|
|
|
|
2012-06-06 10:45:07 +01:00
|
|
|
struct sockaddr_in* in = (struct sockaddr_in*) sockaddr;
|
|
|
|
struct sockaddr_in6* in6 = (struct sockaddr_in6*) sockaddr;
|
|
|
|
|
2012-06-11 14:34:17 +01:00
|
|
|
if (sockaddr->sa_family == AF_INET) {
|
2012-06-06 10:45:07 +01:00
|
|
|
return &in->sin_addr;
|
2012-06-11 14:34:17 +01:00
|
|
|
}
|
|
|
|
if (sockaddr->sa_family == AF_INET6) {
|
2012-06-06 10:45:07 +01:00
|
|
|
return &in6->sin6_addr;
|
2012-06-11 14:34:17 +01:00
|
|
|
}
|
2012-06-06 10:45:07 +01:00
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2012-06-08 18:03:41 +01:00
|
|
|
struct server * server_create (
|
2012-06-27 15:45:33 +01:00
|
|
|
struct flexnbd * flexnbd,
|
2012-06-08 18:03:41 +01:00
|
|
|
char* s_ip_address,
|
|
|
|
char* s_port,
|
|
|
|
char* s_file,
|
|
|
|
int default_deny,
|
|
|
|
int acl_entries,
|
2012-06-21 17:22:34 +01:00
|
|
|
char** s_acl_entries,
|
2012-06-21 18:01:50 +01:00
|
|
|
int max_nbd_clients,
|
|
|
|
int has_control)
|
2012-06-08 18:03:41 +01:00
|
|
|
{
|
2012-06-27 15:45:33 +01:00
|
|
|
NULLCHECK( flexnbd );
|
2012-06-08 18:03:41 +01:00
|
|
|
struct server * out;
|
|
|
|
out = xmalloc( sizeof( struct server ) );
|
2012-06-27 15:45:33 +01:00
|
|
|
out->flexnbd = flexnbd;
|
2012-06-21 18:01:50 +01:00
|
|
|
out->has_control = has_control;
|
2012-06-21 17:22:34 +01:00
|
|
|
out->max_nbd_clients = max_nbd_clients;
|
|
|
|
out->nbd_client = xmalloc( max_nbd_clients * sizeof( struct client_tbl_entry ) );
|
2012-06-08 18:03:41 +01:00
|
|
|
|
|
|
|
out->tcp_backlog = 10; /* does this need to be settable? */
|
|
|
|
|
2012-06-11 13:57:03 +01:00
|
|
|
FATAL_IF_NULL(s_ip_address, "No IP address supplied");
|
|
|
|
FATAL_IF_NULL(s_port, "No port number supplied");
|
|
|
|
FATAL_IF_NULL(s_file, "No filename supplied");
|
2012-06-21 18:01:50 +01:00
|
|
|
NULLCHECK( s_ip_address );
|
2012-06-11 13:57:03 +01:00
|
|
|
FATAL_IF_ZERO(
|
|
|
|
parse_ip_to_sockaddr(&out->bind_to.generic, s_ip_address),
|
|
|
|
"Couldn't parse server address '%s' (use 0 if "
|
|
|
|
"you want to bind to all IPs)",
|
|
|
|
s_ip_address
|
|
|
|
);
|
2012-06-08 18:03:41 +01:00
|
|
|
|
|
|
|
|
|
|
|
out->acl = acl_create( acl_entries, s_acl_entries, default_deny );
|
2012-06-11 14:34:17 +01:00
|
|
|
if (out->acl && out->acl->len != acl_entries) {
|
2012-06-09 02:37:23 +01:00
|
|
|
fatal("Bad ACL entry '%s'", s_acl_entries[out->acl->len]);
|
2012-06-11 14:34:17 +01:00
|
|
|
}
|
2012-06-08 18:03:41 +01:00
|
|
|
|
2012-06-11 13:57:03 +01:00
|
|
|
parse_port( s_port, &out->bind_to.v4 );
|
2012-06-08 18:03:41 +01:00
|
|
|
|
|
|
|
out->filename = s_file;
|
|
|
|
out->filename_incomplete = xmalloc(strlen(s_file)+11+1);
|
|
|
|
strcpy(out->filename_incomplete, s_file);
|
|
|
|
strcpy(out->filename_incomplete + strlen(s_file), ".INCOMPLETE");
|
|
|
|
|
2012-07-11 09:43:16 +01:00
|
|
|
out->l_io = flexthread_mutex_create();
|
|
|
|
out->l_acl= flexthread_mutex_create();
|
2012-06-08 18:03:41 +01:00
|
|
|
|
|
|
|
out->close_signal = self_pipe_create();
|
|
|
|
out->acl_updated_signal = self_pipe_create();
|
|
|
|
|
|
|
|
NULLCHECK( out->close_signal );
|
|
|
|
NULLCHECK( out->acl_updated_signal );
|
|
|
|
|
|
|
|
return out;
|
|
|
|
}
|
|
|
|
|
|
|
|
void server_destroy( struct server * serve )
|
|
|
|
{
|
|
|
|
self_pipe_destroy( serve->acl_updated_signal );
|
2012-06-21 14:15:58 +01:00
|
|
|
serve->acl_updated_signal = NULL;
|
2012-06-08 18:03:41 +01:00
|
|
|
self_pipe_destroy( serve->close_signal );
|
2012-06-21 14:15:58 +01:00
|
|
|
serve->close_signal = NULL;
|
2012-06-08 18:03:41 +01:00
|
|
|
|
2012-07-11 09:43:16 +01:00
|
|
|
flexthread_mutex_destroy( serve->l_acl );
|
|
|
|
flexthread_mutex_destroy( serve->l_io );
|
2012-06-08 18:03:41 +01:00
|
|
|
|
2012-06-21 14:15:58 +01:00
|
|
|
if ( serve->acl ) {
|
|
|
|
acl_destroy( serve->acl );
|
|
|
|
serve->acl = NULL;
|
|
|
|
}
|
2012-06-08 18:03:41 +01:00
|
|
|
|
2012-06-21 15:58:32 +01:00
|
|
|
free( serve->filename_incomplete );
|
|
|
|
|
2012-06-21 17:22:34 +01:00
|
|
|
free( serve->nbd_client );
|
2012-06-08 18:03:41 +01:00
|
|
|
free( serve );
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2012-06-06 11:27:52 +01:00
|
|
|
void server_dirty(struct server *serve, off64_t from, int len)
|
2012-05-21 04:03:17 +01:00
|
|
|
{
|
2012-06-07 11:44:19 +01:00
|
|
|
NULLCHECK( serve );
|
|
|
|
|
2012-06-11 14:34:17 +01:00
|
|
|
if (serve->mirror) {
|
2012-05-21 04:03:17 +01:00
|
|
|
bitset_set_range(serve->mirror->dirty_map, from, len);
|
2012-06-11 14:34:17 +01:00
|
|
|
}
|
2012-05-21 04:03:17 +01:00
|
|
|
}
|
|
|
|
|
2012-06-08 11:02:40 +01:00
|
|
|
#define SERVER_LOCK( s, f, msg ) \
|
2012-06-11 16:08:19 +01:00
|
|
|
do { NULLCHECK( s ); \
|
2012-07-11 09:43:16 +01:00
|
|
|
FATAL_IF( 0 != flexthread_mutex_lock( s->f ), msg ); } while (0)
|
2012-06-08 11:02:40 +01:00
|
|
|
#define SERVER_UNLOCK( s, f, msg ) \
|
2012-06-11 16:08:19 +01:00
|
|
|
do { NULLCHECK( s ); \
|
2012-07-11 09:43:16 +01:00
|
|
|
FATAL_IF( 0 != flexthread_mutex_unlock( s->f ), msg ); } while (0)
|
2012-06-08 11:02:40 +01:00
|
|
|
|
|
|
|
void server_lock_io( struct server * serve)
|
2012-06-06 13:29:13 +01:00
|
|
|
{
|
2012-07-11 09:43:16 +01:00
|
|
|
debug("IO locking");
|
|
|
|
|
2012-06-08 11:02:40 +01:00
|
|
|
SERVER_LOCK( serve, l_io, "Problem with I/O lock" );
|
2012-06-06 13:29:13 +01:00
|
|
|
}
|
2012-06-07 11:44:19 +01:00
|
|
|
|
2012-06-06 13:29:13 +01:00
|
|
|
void server_unlock_io( struct server* serve )
|
|
|
|
{
|
2012-07-11 09:43:16 +01:00
|
|
|
debug("IO unlocking");
|
|
|
|
|
2012-06-08 11:02:40 +01:00
|
|
|
SERVER_UNLOCK( serve, l_io, "Problem with I/O unlock" );
|
2012-06-06 13:29:13 +01:00
|
|
|
}
|
|
|
|
|
2012-07-11 09:43:16 +01:00
|
|
|
|
|
|
|
/* This is only to be called from error handlers. */
|
|
|
|
int server_io_locked( struct server * serve )
|
|
|
|
{
|
|
|
|
NULLCHECK( serve );
|
|
|
|
return flexthread_mutex_held( serve->l_io );
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
2012-06-08 11:02:40 +01:00
|
|
|
void server_lock_acl( struct server *serve )
|
|
|
|
{
|
2012-07-11 09:43:16 +01:00
|
|
|
debug("ACL locking");
|
|
|
|
|
2012-06-08 11:02:40 +01:00
|
|
|
SERVER_LOCK( serve, l_acl, "Problem with ACL lock" );
|
2012-06-06 13:29:13 +01:00
|
|
|
}
|
|
|
|
|
2012-06-08 11:02:40 +01:00
|
|
|
void server_unlock_acl( struct server *serve )
|
2012-06-06 13:29:13 +01:00
|
|
|
{
|
2012-06-08 11:02:40 +01:00
|
|
|
SERVER_UNLOCK( serve, l_acl, "Problem with ACL unlock" );
|
|
|
|
}
|
2012-06-07 11:44:19 +01:00
|
|
|
|
2012-06-08 11:02:40 +01:00
|
|
|
|
2012-07-11 09:43:16 +01:00
|
|
|
int server_acl_locked( struct server * serve )
|
|
|
|
{
|
|
|
|
NULLCHECK( serve );
|
|
|
|
return flexthread_mutex_held( serve->l_acl );
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2012-06-08 18:03:41 +01:00
|
|
|
/** Return the actual port the server bound to. This is used because we
|
|
|
|
* are allowed to pass "0" on the command-line.
|
|
|
|
*/
|
|
|
|
int server_port( struct server * server )
|
|
|
|
{
|
|
|
|
NULLCHECK( server );
|
|
|
|
union mysockaddr addr;
|
|
|
|
socklen_t len = sizeof( addr.v4 );
|
|
|
|
|
|
|
|
if ( getsockname( server->server_fd, &addr.v4, &len ) < 0 ) {
|
2012-06-09 02:37:23 +01:00
|
|
|
fatal( "Failed to get the port number." );
|
2012-06-08 18:03:41 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
return be16toh( addr.v4.sin_port );
|
2012-06-06 13:29:13 +01:00
|
|
|
}
|
|
|
|
|
2012-06-08 18:03:41 +01:00
|
|
|
|
2012-07-12 14:14:46 +01:00
|
|
|
/* Try to bind to our serving socket, retrying until it works or gives a
|
|
|
|
* fatal error. */
|
|
|
|
void serve_bind( struct server * serve )
|
|
|
|
{
|
|
|
|
int bind_result;
|
|
|
|
|
|
|
|
char s_address[64];
|
|
|
|
memset( s_address, 0, 64 );
|
|
|
|
strcpy( s_address, "???" );
|
|
|
|
inet_ntop( serve->bind_to.generic.sa_family,
|
|
|
|
sockaddr_address_data( &serve->bind_to.generic),
|
|
|
|
s_address, 64 );
|
|
|
|
|
|
|
|
do {
|
|
|
|
bind_result = bind(
|
|
|
|
serve->server_fd,
|
|
|
|
&serve->bind_to.generic,
|
|
|
|
sizeof(serve->bind_to));
|
|
|
|
|
|
|
|
if ( 0 == bind_result ) {
|
|
|
|
info( "Bound to %s port %d",
|
|
|
|
s_address,
|
|
|
|
ntohs(serve->bind_to.v4.sin_port));
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
|
|
|
|
warn( "Couldn't bind to %s port %d: %s",
|
|
|
|
s_address,
|
|
|
|
ntohs(serve->bind_to.v4.sin_port),
|
|
|
|
strerror( errno ) );
|
|
|
|
|
|
|
|
switch (errno){
|
|
|
|
/* bind() can give us EACCES,
|
|
|
|
* EADDRINUSE, EADDRNOTAVAIL, EBADF,
|
|
|
|
* EINVAL or ENOTSOCK.
|
|
|
|
*
|
|
|
|
* Any of these other than EACCES,
|
|
|
|
* EADDRINUSE or EADDRNOTAVAIL signify
|
|
|
|
* that there's a logic error somewhere.
|
|
|
|
*/
|
|
|
|
case EACCES:
|
|
|
|
case EADDRINUSE:
|
|
|
|
case EADDRNOTAVAIL:
|
|
|
|
debug("retrying");
|
|
|
|
sleep(1);
|
|
|
|
continue;
|
|
|
|
default:
|
|
|
|
fatal( "Giving up" );
|
|
|
|
}
|
|
|
|
}
|
|
|
|
} while ( 1 );
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
2012-05-29 04:03:28 +01:00
|
|
|
/** Prepares a listening socket for the NBD server, binding etc. */
|
2012-06-06 10:35:50 +01:00
|
|
|
void serve_open_server_socket(struct server* params)
|
2012-05-17 20:14:22 +01:00
|
|
|
{
|
2012-06-07 11:44:19 +01:00
|
|
|
NULLCHECK( params );
|
|
|
|
|
2012-05-29 00:59:12 +01:00
|
|
|
int optval=1;
|
|
|
|
|
2012-06-06 10:19:45 +01:00
|
|
|
params->server_fd= socket(params->bind_to.generic.sa_family == AF_INET ?
|
2012-05-27 14:40:16 +01:00
|
|
|
PF_INET : PF_INET6, SOCK_STREAM, 0);
|
2012-05-17 20:14:22 +01:00
|
|
|
|
2012-06-09 02:25:12 +01:00
|
|
|
FATAL_IF_NEGATIVE(params->server_fd,
|
2012-05-17 20:14:22 +01:00
|
|
|
"Couldn't create server socket");
|
2012-05-27 14:40:16 +01:00
|
|
|
|
2012-07-12 14:14:46 +01:00
|
|
|
/* We need SO_REUSEADDR so that when we switch from listening to
|
|
|
|
* serving we don't have to change address if we don't want to.
|
|
|
|
*
|
|
|
|
* If this fails, it's not necessarily bad in principle, but at
|
|
|
|
* this point in the code we can't tell if it's going to be a
|
|
|
|
* problem. It's also indicative of something odd going on, so
|
|
|
|
* we barf.
|
|
|
|
*/
|
2012-06-09 02:25:12 +01:00
|
|
|
FATAL_IF_NEGATIVE(
|
2012-06-06 10:19:45 +01:00
|
|
|
setsockopt(params->server_fd, SOL_SOCKET, SO_REUSEADDR, &optval, sizeof(optval)),
|
2012-05-29 00:59:12 +01:00
|
|
|
"Couldn't set SO_REUSEADDR"
|
|
|
|
);
|
|
|
|
|
2012-07-12 14:14:46 +01:00
|
|
|
/* TCP_NODELAY makes everything not be slow. If we can't set
|
|
|
|
* this, again, there's something odd going on which we don't
|
|
|
|
* understand.
|
|
|
|
*/
|
2012-06-09 02:25:12 +01:00
|
|
|
FATAL_IF_NEGATIVE(
|
2012-06-06 10:19:45 +01:00
|
|
|
setsockopt(params->server_fd, IPPROTO_TCP, TCP_NODELAY, &optval, sizeof(optval)),
|
2012-05-31 11:33:31 +01:00
|
|
|
"Couldn't set TCP_NODELAY"
|
|
|
|
);
|
|
|
|
|
2012-07-12 14:14:46 +01:00
|
|
|
/* If we can't bind, presumably that's because someone else is
|
|
|
|
* squatting on our ip/port combo, or the ip isn't yet
|
|
|
|
* configured. Ideally we want to retry this. */
|
|
|
|
serve_bind(params);
|
2012-05-17 20:14:22 +01:00
|
|
|
|
2012-06-09 02:25:12 +01:00
|
|
|
FATAL_IF_NEGATIVE(
|
2012-06-06 10:19:45 +01:00
|
|
|
listen(params->server_fd, params->tcp_backlog),
|
2012-05-17 20:14:22 +01:00
|
|
|
"Couldn't listen on server socket"
|
|
|
|
);
|
|
|
|
}
|
|
|
|
|
2012-06-08 18:03:41 +01:00
|
|
|
|
|
|
|
|
2012-06-07 14:25:30 +01:00
|
|
|
int tryjoin_client_thread( struct client_tbl_entry *entry, int (*joinfunc)(pthread_t, void **) )
|
2012-06-07 11:44:19 +01:00
|
|
|
{
|
2012-06-07 14:25:30 +01:00
|
|
|
|
2012-06-07 11:44:19 +01:00
|
|
|
NULLCHECK( entry );
|
2012-06-07 14:25:30 +01:00
|
|
|
NULLCHECK( joinfunc );
|
2012-06-07 11:44:19 +01:00
|
|
|
|
|
|
|
int was_closed = 0;
|
2012-06-21 17:11:12 +01:00
|
|
|
void * status=NULL;
|
2012-06-12 15:08:07 +01:00
|
|
|
int join_errno;
|
2012-06-07 11:44:19 +01:00
|
|
|
|
|
|
|
if (entry->thread != 0) {
|
|
|
|
char s_client_address[64];
|
|
|
|
|
|
|
|
memset(s_client_address, 0, 64);
|
|
|
|
strcpy(s_client_address, "???");
|
|
|
|
inet_ntop( entry->address.generic.sa_family,
|
|
|
|
sockaddr_address_data(&entry->address.generic),
|
|
|
|
s_client_address,
|
|
|
|
64 );
|
|
|
|
|
2012-06-27 15:45:33 +01:00
|
|
|
debug( "%s(%p,...)", joinfunc == pthread_join ? "joining" : "tryjoining", entry->thread );
|
2012-06-12 15:08:07 +01:00
|
|
|
join_errno = joinfunc(entry->thread, &status);
|
|
|
|
/* join_errno can legitimately be ESRCH if the thread is
|
2012-06-21 17:11:12 +01:00
|
|
|
* already dead, but the client still needs tidying up. */
|
2012-06-12 15:08:07 +01:00
|
|
|
if (join_errno != 0 && !entry->client->stopped ) {
|
|
|
|
FATAL_UNLESS( join_errno == EBUSY,
|
|
|
|
"Problem with joining thread %p: %s",
|
|
|
|
entry->thread,
|
|
|
|
strerror(join_errno) );
|
2012-06-07 11:44:19 +01:00
|
|
|
}
|
|
|
|
else {
|
2012-06-22 10:05:41 +01:00
|
|
|
debug("nbd thread %016x exited (%s) with status %ld",
|
|
|
|
entry->thread,
|
2012-06-07 11:44:19 +01:00
|
|
|
s_client_address,
|
|
|
|
(uint64_t)status);
|
|
|
|
client_destroy( entry->client );
|
2012-06-27 15:45:33 +01:00
|
|
|
entry->client = NULL;
|
2012-06-07 11:44:19 +01:00
|
|
|
entry->thread = 0;
|
|
|
|
was_closed = 1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return was_closed;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2012-06-07 14:25:30 +01:00
|
|
|
/**
|
|
|
|
* Check to see if a client thread has finished, and if so, tidy up
|
|
|
|
* after it.
|
|
|
|
* Returns 1 if the thread was cleaned up and the slot freed, 0
|
|
|
|
* otherwise.
|
|
|
|
*
|
|
|
|
* It's important that client_destroy gets called in the same thread
|
|
|
|
* which signals the client threads to stop. This avoids the
|
|
|
|
* possibility of sending a stop signal via a signal which has already
|
|
|
|
* been destroyed. However, it means that stopped client threads,
|
|
|
|
* including their signal pipes, won't be cleaned up until the next new
|
|
|
|
* client connection attempt.
|
|
|
|
*/
|
|
|
|
int cleanup_client_thread( struct client_tbl_entry * entry )
|
|
|
|
{
|
|
|
|
return tryjoin_client_thread( entry, pthread_tryjoin_np );
|
|
|
|
}
|
|
|
|
|
2012-06-21 17:22:34 +01:00
|
|
|
void cleanup_client_threads( struct client_tbl_entry * entries, size_t entries_len )
|
2012-06-21 17:11:12 +01:00
|
|
|
{
|
2012-06-21 18:01:50 +01:00
|
|
|
size_t i;
|
2012-06-21 17:22:34 +01:00
|
|
|
for( i = 0; i < entries_len; i++ ) {
|
2012-06-21 17:11:12 +01:00
|
|
|
cleanup_client_thread( &entries[i] );
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2012-06-07 14:25:30 +01:00
|
|
|
|
|
|
|
/**
|
|
|
|
* Join a client thread after having sent a stop signal to it.
|
|
|
|
* This function will not return until pthread_join has returned, so
|
|
|
|
* ensures that the client thread is dead.
|
|
|
|
*/
|
|
|
|
int join_client_thread( struct client_tbl_entry *entry )
|
|
|
|
{
|
|
|
|
return tryjoin_client_thread( entry, pthread_join );
|
|
|
|
}
|
|
|
|
|
2012-05-29 04:03:28 +01:00
|
|
|
/** We can only accommodate MAX_NBD_CLIENTS connections at once. This function
|
|
|
|
* goes through the current list, waits for any threads that have finished
|
|
|
|
* and returns the next slot free (or -1 if there are none).
|
|
|
|
*/
|
2012-06-06 10:35:50 +01:00
|
|
|
int cleanup_and_find_client_slot(struct server* params)
|
2012-05-27 14:40:16 +01:00
|
|
|
{
|
2012-06-07 11:44:19 +01:00
|
|
|
NULLCHECK( params );
|
|
|
|
|
2012-06-21 17:11:12 +01:00
|
|
|
int slot=-1, i;
|
2012-06-07 11:44:19 +01:00
|
|
|
|
2012-06-21 17:22:34 +01:00
|
|
|
cleanup_client_threads( params->nbd_client, params->max_nbd_clients );
|
2012-06-07 11:44:19 +01:00
|
|
|
|
2012-06-21 17:22:34 +01:00
|
|
|
for ( i = 0; i < params->max_nbd_clients; i++ ) {
|
2012-06-21 17:11:12 +01:00
|
|
|
if( params->nbd_client[i].thread == 0 && slot == -1 ){
|
|
|
|
slot = i;
|
2012-06-07 11:44:19 +01:00
|
|
|
break;
|
2012-05-27 14:40:16 +01:00
|
|
|
}
|
|
|
|
}
|
2012-06-07 11:44:19 +01:00
|
|
|
|
2012-05-27 14:40:16 +01:00
|
|
|
return slot;
|
|
|
|
}
|
|
|
|
|
2012-06-07 11:44:19 +01:00
|
|
|
|
2012-06-08 11:02:40 +01:00
|
|
|
/** Check whether the address client_address is allowed or not according
|
|
|
|
* to the current acl. If params->acl is NULL, the result will be 1,
|
|
|
|
* otherwise it will be the result of acl_includes().
|
|
|
|
*/
|
2012-06-07 11:44:19 +01:00
|
|
|
int server_acl_accepts( struct server *params, union mysockaddr * client_address )
|
|
|
|
{
|
|
|
|
NULLCHECK( params );
|
|
|
|
NULLCHECK( client_address );
|
|
|
|
|
2012-06-08 11:02:40 +01:00
|
|
|
struct acl * acl;
|
|
|
|
int accepted;
|
|
|
|
|
|
|
|
server_lock_acl( params );
|
|
|
|
{
|
|
|
|
acl = params->acl;
|
|
|
|
accepted = acl ? acl_includes( acl, client_address ) : 1;
|
2012-06-07 11:44:19 +01:00
|
|
|
}
|
2012-06-08 11:02:40 +01:00
|
|
|
server_unlock_acl( params );
|
2012-06-07 17:47:43 +01:00
|
|
|
|
2012-06-08 11:02:40 +01:00
|
|
|
return accepted;
|
2012-06-07 11:44:19 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
int server_should_accept_client(
|
|
|
|
struct server * params,
|
|
|
|
union mysockaddr * client_address,
|
|
|
|
char *s_client_address,
|
|
|
|
size_t s_client_address_len )
|
|
|
|
{
|
|
|
|
NULLCHECK( params );
|
|
|
|
NULLCHECK( client_address );
|
|
|
|
NULLCHECK( s_client_address );
|
|
|
|
|
|
|
|
if (inet_ntop(client_address->generic.sa_family,
|
|
|
|
sockaddr_address_data(&client_address->generic),
|
|
|
|
s_client_address, s_client_address_len ) == NULL) {
|
2012-06-22 10:05:41 +01:00
|
|
|
warn( "Rejecting client %s: Bad client_address", s_client_address );
|
2012-06-07 11:44:19 +01:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
if ( !server_acl_accepts( params, client_address ) ) {
|
2012-06-22 10:05:41 +01:00
|
|
|
warn( "Rejecting client %s: Access control error", s_client_address );
|
2012-06-07 11:44:19 +01:00
|
|
|
debug( "We %s have an acl, and default_deny is %s",
|
|
|
|
(params->acl ? "do" : "do not"),
|
2012-06-09 02:25:12 +01:00
|
|
|
(params->acl->default_deny ? "true" : "false") );
|
2012-06-07 11:44:19 +01:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2012-06-21 17:11:12 +01:00
|
|
|
|
|
|
|
int spawn_client_thread(
|
|
|
|
struct client * client_params,
|
|
|
|
pthread_t *out_thread)
|
|
|
|
{
|
2012-06-27 15:45:33 +01:00
|
|
|
int result = pthread_create(out_thread, NULL, client_serve, client_params);
|
2012-06-21 17:11:12 +01:00
|
|
|
|
|
|
|
return result;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2012-05-29 04:03:28 +01:00
|
|
|
/** Dispatch function for accepting an NBD connection and starting a thread
|
|
|
|
* to handle it. Rejects the connection if there is an ACL, and the far end's
|
|
|
|
* address doesn't match, or if there are too many clients already connected.
|
|
|
|
*/
|
2012-06-01 16:24:50 +01:00
|
|
|
void accept_nbd_client(
|
2012-06-06 10:35:50 +01:00
|
|
|
struct server* params,
|
2012-06-01 16:24:50 +01:00
|
|
|
int client_fd,
|
|
|
|
union mysockaddr* client_address)
|
2012-05-18 23:39:16 +01:00
|
|
|
{
|
2012-06-07 11:44:19 +01:00
|
|
|
NULLCHECK(params);
|
|
|
|
NULLCHECK(client_address);
|
|
|
|
|
2012-06-06 11:33:17 +01:00
|
|
|
struct client* client_params;
|
2012-06-07 14:25:30 +01:00
|
|
|
int slot;
|
2012-06-07 11:44:19 +01:00
|
|
|
char s_client_address[64] = {0};
|
2012-06-01 14:48:34 +01:00
|
|
|
|
|
|
|
|
2012-06-22 10:05:41 +01:00
|
|
|
if ( !server_should_accept_client( params, client_address, s_client_address, 64 ) ) {
|
2012-06-07 11:44:19 +01:00
|
|
|
close( client_fd );
|
|
|
|
return;
|
2012-06-01 14:48:34 +01:00
|
|
|
}
|
|
|
|
|
2012-06-07 14:25:30 +01:00
|
|
|
slot = cleanup_and_find_client_slot(params);
|
2012-05-27 14:40:16 +01:00
|
|
|
if (slot < 0) {
|
2012-06-09 02:25:12 +01:00
|
|
|
warn("too many clients to accept connection");
|
2012-05-27 14:40:16 +01:00
|
|
|
close(client_fd);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2012-07-16 11:38:01 +01:00
|
|
|
info( "Client %s accepted.", s_client_address );
|
2012-06-07 11:44:19 +01:00
|
|
|
client_params = client_create( params, client_fd );
|
2012-06-07 14:25:30 +01:00
|
|
|
|
2012-06-07 11:44:19 +01:00
|
|
|
params->nbd_client[slot].client = client_params;
|
2012-06-07 14:25:30 +01:00
|
|
|
memcpy(¶ms->nbd_client[slot].address, client_address,
|
|
|
|
sizeof(union mysockaddr));
|
2012-05-18 23:39:16 +01:00
|
|
|
|
2012-06-21 17:11:12 +01:00
|
|
|
pthread_t * thread = ¶ms->nbd_client[slot].thread;
|
|
|
|
|
2012-06-27 15:45:33 +01:00
|
|
|
if ( 0 != spawn_client_thread( client_params, thread ) ) {
|
2012-06-07 11:44:19 +01:00
|
|
|
debug( "Thread creation problem." );
|
|
|
|
client_destroy( client_params );
|
2012-05-27 14:40:16 +01:00
|
|
|
close(client_fd);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2012-06-12 15:08:07 +01:00
|
|
|
debug("nbd thread %p started (%s)", params->nbd_client[slot].thread, s_client_address);
|
2012-05-18 23:39:16 +01:00
|
|
|
}
|
|
|
|
|
2012-06-06 11:27:52 +01:00
|
|
|
|
2012-06-08 18:03:41 +01:00
|
|
|
void server_audit_clients( struct server * serve)
|
|
|
|
{
|
|
|
|
NULLCHECK( serve );
|
|
|
|
|
|
|
|
int i;
|
|
|
|
struct client_tbl_entry * entry;
|
|
|
|
|
|
|
|
/* There's an apparent race here. If the acl updates while
|
|
|
|
* we're traversing the nbd_clients array, the earlier entries
|
|
|
|
* won't have been audited against the later acl. This isn't a
|
|
|
|
* problem though, because in order to update the acl
|
|
|
|
* server_replace_acl must have been called, so the
|
2012-06-22 10:05:41 +01:00
|
|
|
* server_accept ioop will see a second acl_updated signal as
|
2012-06-08 18:03:41 +01:00
|
|
|
* soon as it hits select, and a second audit will be run.
|
|
|
|
*/
|
2012-06-21 17:22:34 +01:00
|
|
|
for( i = 0; i < serve->max_nbd_clients; i++ ) {
|
2012-06-08 18:03:41 +01:00
|
|
|
entry = &serve->nbd_client[i];
|
|
|
|
if ( 0 == entry->thread ) { continue; }
|
|
|
|
if ( server_acl_accepts( serve, &entry->address ) ) { continue; }
|
|
|
|
client_signal_stop( entry->client );
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2012-06-06 14:25:35 +01:00
|
|
|
int server_is_closed(struct server* serve)
|
2012-06-06 11:27:52 +01:00
|
|
|
{
|
2012-06-07 11:44:19 +01:00
|
|
|
NULLCHECK( serve );
|
|
|
|
return fd_is_closed( serve->server_fd );
|
2012-06-06 11:27:52 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2012-06-07 14:25:30 +01:00
|
|
|
void server_close_clients( struct server *params )
|
|
|
|
{
|
|
|
|
NULLCHECK(params);
|
2012-06-09 02:25:12 +01:00
|
|
|
|
|
|
|
info("closing all clients");
|
2012-06-07 14:25:30 +01:00
|
|
|
|
|
|
|
int i, j;
|
|
|
|
struct client_tbl_entry *entry;
|
|
|
|
|
2012-06-21 17:22:34 +01:00
|
|
|
for( i = 0; i < params->max_nbd_clients; i++ ) {
|
2012-06-07 14:25:30 +01:00
|
|
|
entry = ¶ms->nbd_client[i];
|
|
|
|
|
|
|
|
if ( entry->thread != 0 ) {
|
2012-06-27 15:45:33 +01:00
|
|
|
debug( "Stop signaling client %p", entry->client );
|
2012-06-07 14:25:30 +01:00
|
|
|
client_signal_stop( entry->client );
|
|
|
|
}
|
|
|
|
}
|
2012-06-21 17:22:34 +01:00
|
|
|
for( j = 0; j < params->max_nbd_clients; j++ ) {
|
2012-06-12 15:08:07 +01:00
|
|
|
join_client_thread( ¶ms->nbd_client[j] );
|
2012-06-07 14:25:30 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2012-06-08 18:03:41 +01:00
|
|
|
/** Replace the current acl with a new one. The old one will be thrown
|
|
|
|
* away.
|
|
|
|
*/
|
2012-06-08 10:32:33 +01:00
|
|
|
void server_replace_acl( struct server *serve, struct acl * new_acl )
|
|
|
|
{
|
|
|
|
NULLCHECK(serve);
|
|
|
|
NULLCHECK(new_acl);
|
|
|
|
|
2012-06-08 11:02:40 +01:00
|
|
|
/* We need to lock around updates to the acl in case we try to
|
|
|
|
* destroy the old acl while checking against it.
|
|
|
|
*/
|
|
|
|
server_lock_acl( serve );
|
|
|
|
{
|
|
|
|
struct acl * old_acl = serve->acl;
|
|
|
|
serve->acl = new_acl;
|
|
|
|
/* We should always have an old_acl, but just in case... */
|
|
|
|
if ( old_acl ) { acl_destroy( old_acl ); }
|
|
|
|
}
|
|
|
|
server_unlock_acl( serve );
|
2012-06-08 10:32:33 +01:00
|
|
|
|
|
|
|
self_pipe_signal( serve->acl_updated_signal );
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2012-06-21 17:11:12 +01:00
|
|
|
|
2012-05-29 04:03:28 +01:00
|
|
|
/** Accept either an NBD or control socket connection, dispatch appropriately */
|
2012-06-08 18:03:41 +01:00
|
|
|
int server_accept( struct server * params )
|
2012-05-17 20:14:22 +01:00
|
|
|
{
|
2012-06-07 11:44:19 +01:00
|
|
|
NULLCHECK( params );
|
2012-06-22 10:05:41 +01:00
|
|
|
debug("accept loop starting");
|
2012-06-11 13:57:03 +01:00
|
|
|
int client_fd;
|
2012-06-08 18:03:41 +01:00
|
|
|
union mysockaddr client_address;
|
|
|
|
fd_set fds;
|
|
|
|
socklen_t socklen=sizeof(client_address);
|
2012-06-27 16:17:51 +01:00
|
|
|
/* We select on this fd to receive OS signals (only a few of
|
|
|
|
* which we're interested in, see flexnbd.c */
|
|
|
|
int signal_fd = flexnbd_signal_fd( params->flexnbd );
|
2012-06-08 18:03:41 +01:00
|
|
|
|
|
|
|
FD_ZERO(&fds);
|
|
|
|
FD_SET(params->server_fd, &fds);
|
2012-06-27 16:17:51 +01:00
|
|
|
if( 0 < signal_fd ) { FD_SET(signal_fd, &fds); }
|
2012-06-08 18:03:41 +01:00
|
|
|
self_pipe_fd_set( params->close_signal, &fds );
|
|
|
|
self_pipe_fd_set( params->acl_updated_signal, &fds );
|
|
|
|
|
2012-06-09 02:37:23 +01:00
|
|
|
FATAL_IF_NEGATIVE(select(FD_SETSIZE, &fds,
|
2012-06-08 18:03:41 +01:00
|
|
|
NULL, NULL, NULL), "select() failed");
|
|
|
|
|
|
|
|
if ( self_pipe_fd_isset( params->close_signal, &fds ) ){
|
|
|
|
server_close_clients( params );
|
|
|
|
return 0;
|
2012-05-17 20:14:22 +01:00
|
|
|
}
|
2012-06-08 18:03:41 +01:00
|
|
|
|
2012-06-27 16:17:51 +01:00
|
|
|
if ( 0 < signal_fd && FD_ISSET( signal_fd, &fds ) ){
|
2012-06-27 15:45:33 +01:00
|
|
|
debug( "Stop signal received." );
|
|
|
|
server_close_clients( params );
|
|
|
|
return 0;
|
2012-06-21 17:11:12 +01:00
|
|
|
}
|
|
|
|
|
2012-06-27 15:45:33 +01:00
|
|
|
|
2012-06-08 18:03:41 +01:00
|
|
|
if ( self_pipe_fd_isset( params->acl_updated_signal, &fds ) ) {
|
2012-06-11 13:49:35 +01:00
|
|
|
self_pipe_signal_clear( params->acl_updated_signal );
|
2012-06-08 18:03:41 +01:00
|
|
|
server_audit_clients( params );
|
|
|
|
}
|
|
|
|
|
2012-06-11 13:49:35 +01:00
|
|
|
if ( FD_ISSET( params->server_fd, &fds ) ){
|
|
|
|
client_fd = accept( params->server_fd, &client_address.generic, &socklen );
|
2012-06-08 18:03:41 +01:00
|
|
|
debug("Accepted nbd client socket");
|
|
|
|
accept_nbd_client(params, client_fd, &client_address);
|
2012-06-21 18:01:50 +01:00
|
|
|
}
|
2012-06-11 13:49:35 +01:00
|
|
|
|
2012-06-08 18:03:41 +01:00
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
void serve_accept_loop(struct server* params)
|
|
|
|
{
|
2012-06-22 10:05:41 +01:00
|
|
|
NULLCHECK( params );
|
2012-06-08 18:03:41 +01:00
|
|
|
while( server_accept( params ) );
|
2012-05-17 20:14:22 +01:00
|
|
|
}
|
|
|
|
|
2012-05-29 04:03:28 +01:00
|
|
|
/** Initialisation function that sets up the initial allocation map, i.e. so
|
|
|
|
* we know which blocks of the file are allocated.
|
|
|
|
*/
|
2012-06-06 10:35:50 +01:00
|
|
|
void serve_init_allocation_map(struct server* params)
|
2012-05-18 13:24:35 +01:00
|
|
|
{
|
2012-06-07 11:44:19 +01:00
|
|
|
NULLCHECK( params );
|
|
|
|
|
2012-05-18 13:24:35 +01:00
|
|
|
int fd = open(params->filename, O_RDONLY);
|
|
|
|
off64_t size;
|
2012-06-07 11:44:19 +01:00
|
|
|
|
2012-06-09 02:25:12 +01:00
|
|
|
FATAL_IF_NEGATIVE(fd, "Couldn't open %s", params->filename);
|
2012-05-18 13:24:35 +01:00
|
|
|
size = lseek64(fd, 0, SEEK_END);
|
2012-05-21 04:03:17 +01:00
|
|
|
params->size = size;
|
2012-06-09 02:25:12 +01:00
|
|
|
FATAL_IF_NEGATIVE(size, "Couldn't find size of %s",
|
2012-06-08 18:03:41 +01:00
|
|
|
params->filename);
|
2012-06-07 11:17:02 +01:00
|
|
|
params->allocation_map =
|
2012-05-18 13:24:35 +01:00
|
|
|
build_allocation_map(fd, size, block_allocation_resolution);
|
|
|
|
close(fd);
|
|
|
|
}
|
|
|
|
|
2012-06-06 12:41:03 +01:00
|
|
|
|
|
|
|
/* Tell the server to close all the things. */
|
|
|
|
void serve_signal_close( struct server * serve )
|
|
|
|
{
|
2012-06-07 11:44:19 +01:00
|
|
|
NULLCHECK( serve );
|
2012-06-09 02:25:12 +01:00
|
|
|
info("signalling close");
|
2012-06-06 12:41:03 +01:00
|
|
|
self_pipe_signal( serve->close_signal );
|
|
|
|
}
|
|
|
|
|
2012-06-13 13:44:21 +01:00
|
|
|
/* Block until the server closes the server_fd.
|
|
|
|
*/
|
|
|
|
void serve_wait_for_close( struct server * serve )
|
|
|
|
{
|
|
|
|
while( !fd_is_closed( serve->server_fd ) ){
|
|
|
|
usleep(10000);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2012-06-21 18:01:50 +01:00
|
|
|
/* We've just had an ENTRUST/DISCONNECT pair, so we need to shut down
|
|
|
|
* and signal our listener that we can safely take over.
|
|
|
|
*/
|
|
|
|
void server_control_arrived( struct server *serve )
|
|
|
|
{
|
|
|
|
NULLCHECK( serve );
|
|
|
|
|
|
|
|
serve->has_control = 1;
|
|
|
|
serve_signal_close( serve );
|
|
|
|
}
|
|
|
|
|
2012-06-06 12:41:03 +01:00
|
|
|
|
2012-05-29 04:03:28 +01:00
|
|
|
/** Closes sockets, frees memory and waits for all client threads to finish */
|
2012-06-11 13:57:03 +01:00
|
|
|
void serve_cleanup(struct server* params,
|
|
|
|
int fatal __attribute__ ((unused)) )
|
2012-05-29 04:03:28 +01:00
|
|
|
{
|
2012-06-07 11:44:19 +01:00
|
|
|
NULLCHECK( params );
|
2012-06-09 02:25:12 +01:00
|
|
|
|
|
|
|
info("cleaning up");
|
2012-06-07 11:44:19 +01:00
|
|
|
|
2012-05-29 04:03:28 +01:00
|
|
|
int i;
|
|
|
|
|
2012-06-11 13:57:03 +01:00
|
|
|
if (params->server_fd){ close(params->server_fd); }
|
2012-06-06 12:41:03 +01:00
|
|
|
|
2012-06-11 13:57:03 +01:00
|
|
|
if (params->allocation_map) {
|
2012-06-09 02:25:12 +01:00
|
|
|
free(params->allocation_map);
|
2012-06-11 13:57:03 +01:00
|
|
|
}
|
2012-05-29 04:03:28 +01:00
|
|
|
|
2012-06-27 15:45:33 +01:00
|
|
|
if (params->mirror_super) {
|
|
|
|
/* AWOOGA! RACE! */
|
|
|
|
pthread_t mirror_t = params->mirror_super->thread;
|
2012-06-09 02:25:12 +01:00
|
|
|
params->mirror->signal_abandon = 1;
|
2012-06-27 15:45:33 +01:00
|
|
|
pthread_join( mirror_t, NULL );
|
2012-06-09 02:25:12 +01:00
|
|
|
}
|
2012-05-29 04:03:28 +01:00
|
|
|
|
2012-06-21 17:22:34 +01:00
|
|
|
for (i=0; i < params->max_nbd_clients; i++) {
|
2012-05-29 04:03:28 +01:00
|
|
|
void* status;
|
2012-06-12 15:08:07 +01:00
|
|
|
pthread_t thread_id = params->nbd_client[i].thread;
|
2012-05-29 04:03:28 +01:00
|
|
|
|
2012-06-12 15:08:07 +01:00
|
|
|
if (thread_id != 0) {
|
|
|
|
debug("joining thread %p", thread_id);
|
|
|
|
pthread_join(thread_id, &status);
|
2012-05-29 04:03:28 +01:00
|
|
|
}
|
|
|
|
}
|
2012-07-11 09:43:16 +01:00
|
|
|
|
|
|
|
if ( server_acl_locked( params ) ) {
|
|
|
|
server_unlock_acl( params );
|
|
|
|
}
|
|
|
|
|
2012-06-22 10:05:41 +01:00
|
|
|
debug( "Cleanup done");
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
int server_is_in_control( struct server *serve )
|
|
|
|
{
|
|
|
|
NULLCHECK( serve );
|
|
|
|
return serve->has_control;
|
2012-05-29 04:03:28 +01:00
|
|
|
}
|
|
|
|
|
2012-06-27 15:45:33 +01:00
|
|
|
int server_default_deny( struct server * serve )
|
|
|
|
{
|
|
|
|
NULLCHECK( serve );
|
|
|
|
return acl_default_deny( serve->acl );
|
|
|
|
}
|
2012-06-22 10:05:41 +01:00
|
|
|
|
2012-05-29 04:03:28 +01:00
|
|
|
/** Full lifecycle of the server */
|
2012-06-21 18:01:50 +01:00
|
|
|
int do_serve(struct server* params)
|
2012-05-17 20:14:22 +01:00
|
|
|
{
|
2012-06-07 11:44:19 +01:00
|
|
|
NULLCHECK( params );
|
2012-06-21 18:01:50 +01:00
|
|
|
|
|
|
|
int has_control;
|
2012-06-09 02:25:12 +01:00
|
|
|
|
|
|
|
error_set_handler((cleanup_handler*) serve_cleanup, params);
|
2012-05-18 18:44:34 +01:00
|
|
|
serve_open_server_socket(params);
|
2012-05-18 13:24:35 +01:00
|
|
|
serve_init_allocation_map(params);
|
2012-05-17 20:14:22 +01:00
|
|
|
serve_accept_loop(params);
|
2012-06-21 18:01:50 +01:00
|
|
|
has_control = params->has_control;
|
2012-06-09 02:25:12 +01:00
|
|
|
serve_cleanup(params, 0);
|
2012-06-21 18:01:50 +01:00
|
|
|
|
|
|
|
return has_control;
|
2012-05-17 20:14:22 +01:00
|
|
|
}
|
|
|
|
|