mirror of
https://github.com/fluencelabs/redis
synced 2025-04-01 23:31:03 +00:00
IO performances greatly improved under high writes load
This commit is contained in:
parent
6eaad66373
commit
f771dc23a0
115
src/dscache.c
115
src/dscache.c
@ -107,7 +107,8 @@
|
|||||||
*/
|
*/
|
||||||
|
|
||||||
void spawnIOThread(void);
|
void spawnIOThread(void);
|
||||||
int cacheScheduleIOPushJobs(int onlyloads);
|
int cacheScheduleIOPushJobs(int flags);
|
||||||
|
int processActiveIOJobs(int max);
|
||||||
|
|
||||||
/* =================== Virtual Memory - Blocking Side ====================== */
|
/* =================== Virtual Memory - Blocking Side ====================== */
|
||||||
|
|
||||||
@ -216,9 +217,9 @@ int cacheFreeOneEntry(void) {
|
|||||||
* otherwise we'll use an infinite amount of memory if changes to
|
* otherwise we'll use an infinite amount of memory if changes to
|
||||||
* the dataset are faster than I/O */
|
* the dataset are faster than I/O */
|
||||||
if (listLength(server.cache_io_queue) > 0) {
|
if (listLength(server.cache_io_queue) > 0) {
|
||||||
cacheScheduleIOPushJobs(0);
|
redisLog(REDIS_DEBUG,"--- Busy waiting IO to reclaim memory");
|
||||||
waitEmptyIOJobsQueue();
|
cacheScheduleIOPushJobs(REDIS_IO_ASAP);
|
||||||
processAllPendingIOJobs();
|
processActiveIOJobs(1);
|
||||||
return REDIS_OK;
|
return REDIS_OK;
|
||||||
}
|
}
|
||||||
/* Nothing to free at all... */
|
/* Nothing to free at all... */
|
||||||
@ -484,62 +485,86 @@ void spawnIOThread(void) {
|
|||||||
server.io_active_threads++;
|
server.io_active_threads++;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Wait that all the pending IO Jobs are processed */
|
/* Wait that up to 'max' pending IO Jobs are processed by the I/O thread.
|
||||||
void waitEmptyIOJobsQueue(void) {
|
* From our point of view an IO job processed means that the count of
|
||||||
while(1) {
|
* server.io_processed must increase by one.
|
||||||
|
*
|
||||||
|
* If max is -1, all the pending IO jobs will be processed.
|
||||||
|
*
|
||||||
|
* Returns the number of IO jobs processed.
|
||||||
|
*
|
||||||
|
* NOTE: while this may appear like a busy loop, we are actually blocked
|
||||||
|
* by IO since we continuously acquire/release the IO lock. */
|
||||||
|
int processActiveIOJobs(int max) {
|
||||||
|
int processed = 0;
|
||||||
|
|
||||||
|
while(max == -1 || max > 0) {
|
||||||
int io_processed_len;
|
int io_processed_len;
|
||||||
|
|
||||||
lockThreadedIO();
|
lockThreadedIO();
|
||||||
if (listLength(server.io_newjobs) == 0 &&
|
if (listLength(server.io_newjobs) == 0 &&
|
||||||
listLength(server.io_processing) == 0)
|
listLength(server.io_processing) == 0)
|
||||||
{
|
{
|
||||||
|
/* There is nothing more to process */
|
||||||
unlockThreadedIO();
|
unlockThreadedIO();
|
||||||
return;
|
break;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#if 0
|
||||||
/* If there are new jobs we need to signal the thread to
|
/* If there are new jobs we need to signal the thread to
|
||||||
* process the next one. */
|
* process the next one. */
|
||||||
redisLog(REDIS_DEBUG,"waitEmptyIOJobsQueue: new %d, processing %d",
|
redisLog(REDIS_DEBUG,"waitEmptyIOJobsQueue: new %d, processing %d",
|
||||||
listLength(server.io_newjobs),
|
listLength(server.io_newjobs),
|
||||||
listLength(server.io_processing));
|
listLength(server.io_processing));
|
||||||
|
|
||||||
/* FIXME: signal or not?
|
|
||||||
if (listLength(server.io_newjobs)) {
|
if (listLength(server.io_newjobs)) {
|
||||||
pthread_cond_signal(&server.io_condvar);
|
pthread_cond_signal(&server.io_condvar);
|
||||||
}
|
}
|
||||||
*/
|
#endif
|
||||||
/* While waiting for empty jobs queue condition we post-process some
|
|
||||||
* finshed job, as I/O threads may be hanging trying to write against
|
/* Check if we can process some finished job */
|
||||||
* the io_ready_pipe_write FD but there are so much pending jobs that
|
|
||||||
* it's blocking. */
|
|
||||||
io_processed_len = listLength(server.io_processed);
|
io_processed_len = listLength(server.io_processed);
|
||||||
unlockThreadedIO();
|
unlockThreadedIO();
|
||||||
if (io_processed_len) {
|
if (io_processed_len) {
|
||||||
vmThreadedIOCompletedJob(NULL,server.io_ready_pipe_read,
|
vmThreadedIOCompletedJob(NULL,server.io_ready_pipe_read,
|
||||||
(void*)0xdeadbeef,0);
|
(void*)0xdeadbeef,0);
|
||||||
/* FIXME: probably wiser to drop this sleeps. Anyway
|
processed++;
|
||||||
* the contention on the IO thread will avoid we to loop
|
if (max != -1) max--;
|
||||||
* too fast here. */
|
|
||||||
usleep(1000); /* 1 millisecond */
|
|
||||||
} else {
|
|
||||||
/* FIXME: same as fixme above. */
|
|
||||||
usleep(10000); /* 10 milliseconds */
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
return processed;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Process all the IO Jobs already completed by threads but still waiting
|
void waitEmptyIOJobsQueue(void) {
|
||||||
* processing from the main thread. */
|
processActiveIOJobs(-1);
|
||||||
void processAllPendingIOJobs(void) {
|
}
|
||||||
while(1) {
|
|
||||||
|
/* Process up to 'max' IO Jobs already completed by threads but still waiting
|
||||||
|
* processing from the main thread.
|
||||||
|
*
|
||||||
|
* If max == -1 all the pending jobs are processed.
|
||||||
|
*
|
||||||
|
* The number of processed jobs is returned. */
|
||||||
|
int processPendingIOJobs(int max) {
|
||||||
|
int processed = 0;
|
||||||
|
|
||||||
|
while(max == -1 || max > 0) {
|
||||||
int io_processed_len;
|
int io_processed_len;
|
||||||
|
|
||||||
lockThreadedIO();
|
lockThreadedIO();
|
||||||
io_processed_len = listLength(server.io_processed);
|
io_processed_len = listLength(server.io_processed);
|
||||||
unlockThreadedIO();
|
unlockThreadedIO();
|
||||||
if (io_processed_len == 0) return;
|
if (io_processed_len == 0) break;
|
||||||
vmThreadedIOCompletedJob(NULL,server.io_ready_pipe_read,
|
vmThreadedIOCompletedJob(NULL,server.io_ready_pipe_read,
|
||||||
(void*)0xdeadbeef,0);
|
(void*)0xdeadbeef,0);
|
||||||
|
if (max != -1) max--;
|
||||||
|
processed++;
|
||||||
}
|
}
|
||||||
|
return processed;
|
||||||
|
}
|
||||||
|
|
||||||
|
void processAllPendingIOJobs(void) {
|
||||||
|
processPendingIOJobs(-1);
|
||||||
}
|
}
|
||||||
|
|
||||||
/* This function must be called while with threaded IO locked */
|
/* This function must be called while with threaded IO locked */
|
||||||
@ -665,7 +690,7 @@ void cacheScheduleIO(redisDb *db, robj *key, int type) {
|
|||||||
* in queue for the same key. */
|
* in queue for the same key. */
|
||||||
if (type == REDIS_IO_LOAD && !(flags & REDIS_IO_SAVE)) {
|
if (type == REDIS_IO_LOAD && !(flags & REDIS_IO_SAVE)) {
|
||||||
listAddNodeHead(server.cache_io_queue, op);
|
listAddNodeHead(server.cache_io_queue, op);
|
||||||
cacheScheduleIOPushJobs(1);
|
cacheScheduleIOPushJobs(REDIS_IO_ONLYLOADS);
|
||||||
} else {
|
} else {
|
||||||
/* FIXME: probably when this happens we want to at least move
|
/* FIXME: probably when this happens we want to at least move
|
||||||
* the write job about this queue on top, and set the creation time
|
* the write job about this queue on top, and set the creation time
|
||||||
@ -675,13 +700,19 @@ void cacheScheduleIO(redisDb *db, robj *key, int type) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
/* Push scheduled IO operations into IO Jobs that the IO thread can process.
|
/* Push scheduled IO operations into IO Jobs that the IO thread can process.
|
||||||
* If 'onlyloads' is true only IO_LOAD jobs are processed: this is useful
|
*
|
||||||
* since it's save to push LOAD IO jobs from any place of the code, while
|
* If flags include REDIS_IO_ONLYLOADS only load jobs are processed:this is
|
||||||
|
* useful since it's safe to push LOAD IO jobs from any place of the code, while
|
||||||
* SAVE io jobs should never be pushed while we are processing a command
|
* SAVE io jobs should never be pushed while we are processing a command
|
||||||
* (not protected by lookupKey() that will block on keys in IO_SAVEINPROG
|
* (not protected by lookupKey() that will block on keys in IO_SAVEINPROG
|
||||||
* state. */
|
* state.
|
||||||
|
*
|
||||||
|
* The REDIS_IO_ASAP flag tells the function to don't wait for the IO job
|
||||||
|
* scheduled completion time, but just do the operation ASAP. This is useful
|
||||||
|
* when we need to reclaim memory from the IO queue.
|
||||||
|
*/
|
||||||
#define MAX_IO_JOBS_QUEUE 100
|
#define MAX_IO_JOBS_QUEUE 100
|
||||||
int cacheScheduleIOPushJobs(int onlyloads) {
|
int cacheScheduleIOPushJobs(int flags) {
|
||||||
time_t now = time(NULL);
|
time_t now = time(NULL);
|
||||||
listNode *ln;
|
listNode *ln;
|
||||||
int jobs, topush = 0, pushed = 0;
|
int jobs, topush = 0, pushed = 0;
|
||||||
@ -699,24 +730,25 @@ int cacheScheduleIOPushJobs(int onlyloads) {
|
|||||||
|
|
||||||
while((ln = listFirst(server.cache_io_queue)) != NULL) {
|
while((ln = listFirst(server.cache_io_queue)) != NULL) {
|
||||||
ioop *op = ln->value;
|
ioop *op = ln->value;
|
||||||
|
struct dictEntry *de;
|
||||||
|
robj *val;
|
||||||
|
|
||||||
if (!topush) break;
|
if (!topush) break;
|
||||||
topush--;
|
topush--;
|
||||||
|
|
||||||
if (op->type == REDIS_IO_LOAD ||
|
if (op->type != REDIS_IO_LOAD && flags & REDIS_IO_ONLYLOADS) break;
|
||||||
(!onlyloads && (now - op->ctime) >= server.cache_flush_delay))
|
|
||||||
{
|
|
||||||
struct dictEntry *de;
|
|
||||||
robj *val;
|
|
||||||
|
|
||||||
/* Don't add a SAVE job in queue if there is already
|
if (!(flags & REDIS_IO_ASAP) &&
|
||||||
|
(now - op->ctime) < server.cache_flush_delay) break;
|
||||||
|
|
||||||
|
/* Don't add a SAVE job in the IO thread queue if there is already
|
||||||
* a save in progress for the same key. */
|
* a save in progress for the same key. */
|
||||||
if (op->type == REDIS_IO_SAVE &&
|
if (op->type == REDIS_IO_SAVE &&
|
||||||
cacheScheduleIOGetFlags(op->db,op->key) & REDIS_IO_SAVEINPROG)
|
cacheScheduleIOGetFlags(op->db,op->key) & REDIS_IO_SAVEINPROG)
|
||||||
{
|
{
|
||||||
/* Move the operation at the end of the list of there
|
/* Move the operation at the end of the list if there
|
||||||
* are other operations. Otherwise break, nothing to do
|
* are other operations, so we can try to process the next one.
|
||||||
* here. */
|
* Otherwise break, nothing to do here. */
|
||||||
if (listLength(server.cache_io_queue) > 1) {
|
if (listLength(server.cache_io_queue) > 1) {
|
||||||
listDelNode(server.cache_io_queue,ln);
|
listDelNode(server.cache_io_queue,ln);
|
||||||
listAddNodeTail(server.cache_io_queue,op);
|
listAddNodeTail(server.cache_io_queue,op);
|
||||||
@ -756,9 +788,6 @@ int cacheScheduleIOPushJobs(int onlyloads) {
|
|||||||
decrRefCount(op->key);
|
decrRefCount(op->key);
|
||||||
zfree(op);
|
zfree(op);
|
||||||
pushed++;
|
pushed++;
|
||||||
} else {
|
|
||||||
break; /* too early */
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
return pushed;
|
return pushed;
|
||||||
}
|
}
|
||||||
|
@ -125,6 +125,10 @@
|
|||||||
#define REDIS_IO_LOADINPROG 4
|
#define REDIS_IO_LOADINPROG 4
|
||||||
#define REDIS_IO_SAVEINPROG 8
|
#define REDIS_IO_SAVEINPROG 8
|
||||||
|
|
||||||
|
/* Generic IO flags */
|
||||||
|
#define REDIS_IO_ONLYLOADS 1
|
||||||
|
#define REDIS_IO_ASAP 2
|
||||||
|
|
||||||
#define REDIS_MAX_COMPLETED_JOBS_PROCESSED 1
|
#define REDIS_MAX_COMPLETED_JOBS_PROCESSED 1
|
||||||
#define REDIS_THREAD_STACK_SIZE (1024*1024*4)
|
#define REDIS_THREAD_STACK_SIZE (1024*1024*4)
|
||||||
|
|
||||||
|
Loading…
x
Reference in New Issue
Block a user