IO performances greatly improved under high writes load

This commit is contained in:
antirez 2011-01-05 12:34:14 +01:00
parent 6eaad66373
commit f771dc23a0
2 changed files with 116 additions and 83 deletions

View File

@ -107,7 +107,8 @@
*/ */
void spawnIOThread(void); void spawnIOThread(void);
int cacheScheduleIOPushJobs(int onlyloads); int cacheScheduleIOPushJobs(int flags);
int processActiveIOJobs(int max);
/* =================== Virtual Memory - Blocking Side ====================== */ /* =================== Virtual Memory - Blocking Side ====================== */
@ -216,9 +217,9 @@ int cacheFreeOneEntry(void) {
* otherwise we'll use an infinite amount of memory if changes to * otherwise we'll use an infinite amount of memory if changes to
* the dataset are faster than I/O */ * the dataset are faster than I/O */
if (listLength(server.cache_io_queue) > 0) { if (listLength(server.cache_io_queue) > 0) {
cacheScheduleIOPushJobs(0); redisLog(REDIS_DEBUG,"--- Busy waiting IO to reclaim memory");
waitEmptyIOJobsQueue(); cacheScheduleIOPushJobs(REDIS_IO_ASAP);
processAllPendingIOJobs(); processActiveIOJobs(1);
return REDIS_OK; return REDIS_OK;
} }
/* Nothing to free at all... */ /* Nothing to free at all... */
@ -484,62 +485,86 @@ void spawnIOThread(void) {
server.io_active_threads++; server.io_active_threads++;
} }
/* Wait that all the pending IO Jobs are processed */ /* Wait that up to 'max' pending IO Jobs are processed by the I/O thread.
void waitEmptyIOJobsQueue(void) { * From our point of view an IO job processed means that the count of
while(1) { * server.io_processed must increase by one.
*
* If max is -1, all the pending IO jobs will be processed.
*
* Returns the number of IO jobs processed.
*
* NOTE: while this may appear like a busy loop, we are actually blocked
* by IO since we continuously acquire/release the IO lock. */
int processActiveIOJobs(int max) {
int processed = 0;
while(max == -1 || max > 0) {
int io_processed_len; int io_processed_len;
lockThreadedIO(); lockThreadedIO();
if (listLength(server.io_newjobs) == 0 && if (listLength(server.io_newjobs) == 0 &&
listLength(server.io_processing) == 0) listLength(server.io_processing) == 0)
{ {
/* There is nothing more to process */
unlockThreadedIO(); unlockThreadedIO();
return; break;
} }
#if 0
/* If there are new jobs we need to signal the thread to /* If there are new jobs we need to signal the thread to
* process the next one. */ * process the next one. */
redisLog(REDIS_DEBUG,"waitEmptyIOJobsQueue: new %d, processing %d", redisLog(REDIS_DEBUG,"waitEmptyIOJobsQueue: new %d, processing %d",
listLength(server.io_newjobs), listLength(server.io_newjobs),
listLength(server.io_processing)); listLength(server.io_processing));
/* FIXME: signal or not?
if (listLength(server.io_newjobs)) { if (listLength(server.io_newjobs)) {
pthread_cond_signal(&server.io_condvar); pthread_cond_signal(&server.io_condvar);
} }
*/ #endif
/* While waiting for empty jobs queue condition we post-process some
* finshed job, as I/O threads may be hanging trying to write against /* Check if we can process some finished job */
* the io_ready_pipe_write FD but there are so much pending jobs that
* it's blocking. */
io_processed_len = listLength(server.io_processed); io_processed_len = listLength(server.io_processed);
unlockThreadedIO(); unlockThreadedIO();
if (io_processed_len) { if (io_processed_len) {
vmThreadedIOCompletedJob(NULL,server.io_ready_pipe_read, vmThreadedIOCompletedJob(NULL,server.io_ready_pipe_read,
(void*)0xdeadbeef,0); (void*)0xdeadbeef,0);
/* FIXME: probably wiser to drop this sleeps. Anyway processed++;
* the contention on the IO thread will avoid we to loop if (max != -1) max--;
* too fast here. */
usleep(1000); /* 1 millisecond */
} else {
/* FIXME: same as fixme above. */
usleep(10000); /* 10 milliseconds */
} }
} }
return processed;
} }
/* Process all the IO Jobs already completed by threads but still waiting void waitEmptyIOJobsQueue(void) {
* processing from the main thread. */ processActiveIOJobs(-1);
void processAllPendingIOJobs(void) { }
while(1) {
/* Process up to 'max' IO Jobs already completed by threads but still waiting
* processing from the main thread.
*
* If max == -1 all the pending jobs are processed.
*
* The number of processed jobs is returned. */
int processPendingIOJobs(int max) {
int processed = 0;
while(max == -1 || max > 0) {
int io_processed_len; int io_processed_len;
lockThreadedIO(); lockThreadedIO();
io_processed_len = listLength(server.io_processed); io_processed_len = listLength(server.io_processed);
unlockThreadedIO(); unlockThreadedIO();
if (io_processed_len == 0) return; if (io_processed_len == 0) break;
vmThreadedIOCompletedJob(NULL,server.io_ready_pipe_read, vmThreadedIOCompletedJob(NULL,server.io_ready_pipe_read,
(void*)0xdeadbeef,0); (void*)0xdeadbeef,0);
if (max != -1) max--;
processed++;
} }
return processed;
}
void processAllPendingIOJobs(void) {
processPendingIOJobs(-1);
} }
/* This function must be called while with threaded IO locked */ /* This function must be called while with threaded IO locked */
@ -665,7 +690,7 @@ void cacheScheduleIO(redisDb *db, robj *key, int type) {
* in queue for the same key. */ * in queue for the same key. */
if (type == REDIS_IO_LOAD && !(flags & REDIS_IO_SAVE)) { if (type == REDIS_IO_LOAD && !(flags & REDIS_IO_SAVE)) {
listAddNodeHead(server.cache_io_queue, op); listAddNodeHead(server.cache_io_queue, op);
cacheScheduleIOPushJobs(1); cacheScheduleIOPushJobs(REDIS_IO_ONLYLOADS);
} else { } else {
/* FIXME: probably when this happens we want to at least move /* FIXME: probably when this happens we want to at least move
* the write job about this queue on top, and set the creation time * the write job about this queue on top, and set the creation time
@ -675,13 +700,19 @@ void cacheScheduleIO(redisDb *db, robj *key, int type) {
} }
/* Push scheduled IO operations into IO Jobs that the IO thread can process. /* Push scheduled IO operations into IO Jobs that the IO thread can process.
* If 'onlyloads' is true only IO_LOAD jobs are processed: this is useful *
* since it's save to push LOAD IO jobs from any place of the code, while * If flags include REDIS_IO_ONLYLOADS only load jobs are processed:this is
* useful since it's safe to push LOAD IO jobs from any place of the code, while
* SAVE io jobs should never be pushed while we are processing a command * SAVE io jobs should never be pushed while we are processing a command
* (not protected by lookupKey() that will block on keys in IO_SAVEINPROG * (not protected by lookupKey() that will block on keys in IO_SAVEINPROG
* state. */ * state.
*
* The REDIS_IO_ASAP flag tells the function to don't wait for the IO job
* scheduled completion time, but just do the operation ASAP. This is useful
* when we need to reclaim memory from the IO queue.
*/
#define MAX_IO_JOBS_QUEUE 100 #define MAX_IO_JOBS_QUEUE 100
int cacheScheduleIOPushJobs(int onlyloads) { int cacheScheduleIOPushJobs(int flags) {
time_t now = time(NULL); time_t now = time(NULL);
listNode *ln; listNode *ln;
int jobs, topush = 0, pushed = 0; int jobs, topush = 0, pushed = 0;
@ -699,24 +730,25 @@ int cacheScheduleIOPushJobs(int onlyloads) {
while((ln = listFirst(server.cache_io_queue)) != NULL) { while((ln = listFirst(server.cache_io_queue)) != NULL) {
ioop *op = ln->value; ioop *op = ln->value;
struct dictEntry *de;
robj *val;
if (!topush) break; if (!topush) break;
topush--; topush--;
if (op->type == REDIS_IO_LOAD || if (op->type != REDIS_IO_LOAD && flags & REDIS_IO_ONLYLOADS) break;
(!onlyloads && (now - op->ctime) >= server.cache_flush_delay))
{
struct dictEntry *de;
robj *val;
/* Don't add a SAVE job in queue if there is already if (!(flags & REDIS_IO_ASAP) &&
(now - op->ctime) < server.cache_flush_delay) break;
/* Don't add a SAVE job in the IO thread queue if there is already
* a save in progress for the same key. */ * a save in progress for the same key. */
if (op->type == REDIS_IO_SAVE && if (op->type == REDIS_IO_SAVE &&
cacheScheduleIOGetFlags(op->db,op->key) & REDIS_IO_SAVEINPROG) cacheScheduleIOGetFlags(op->db,op->key) & REDIS_IO_SAVEINPROG)
{ {
/* Move the operation at the end of the list of there /* Move the operation at the end of the list if there
* are other operations. Otherwise break, nothing to do * are other operations, so we can try to process the next one.
* here. */ * Otherwise break, nothing to do here. */
if (listLength(server.cache_io_queue) > 1) { if (listLength(server.cache_io_queue) > 1) {
listDelNode(server.cache_io_queue,ln); listDelNode(server.cache_io_queue,ln);
listAddNodeTail(server.cache_io_queue,op); listAddNodeTail(server.cache_io_queue,op);
@ -756,9 +788,6 @@ int cacheScheduleIOPushJobs(int onlyloads) {
decrRefCount(op->key); decrRefCount(op->key);
zfree(op); zfree(op);
pushed++; pushed++;
} else {
break; /* too early */
}
} }
return pushed; return pushed;
} }

View File

@ -125,6 +125,10 @@
#define REDIS_IO_LOADINPROG 4 #define REDIS_IO_LOADINPROG 4
#define REDIS_IO_SAVEINPROG 8 #define REDIS_IO_SAVEINPROG 8
/* Generic IO flags */
#define REDIS_IO_ONLYLOADS 1
#define REDIS_IO_ASAP 2
#define REDIS_MAX_COMPLETED_JOBS_PROCESSED 1 #define REDIS_MAX_COMPLETED_JOBS_PROCESSED 1
#define REDIS_THREAD_STACK_SIZE (1024*1024*4) #define REDIS_THREAD_STACK_SIZE (1024*1024*4)