summaryrefslogtreecommitdiff
path: root/sys
diff options
context:
space:
mode:
authorBob Beck <beck@cvs.openbsd.org>2012-11-07 17:50:50 +0000
committerBob Beck <beck@cvs.openbsd.org>2012-11-07 17:50:50 +0000
commitb99784fe662faa712f3630b22e11e491a71180a5 (patch)
tree5e2f33011ce5e2868d515542f2ce7177cf0ad387 /sys
parent8c9e508e5ee189a27cad6ae4d3988986f019f80c (diff)
Fix the buffer cache.
A long time ago (in vienna) the reserves for the cleaner and syncer were removed. softdep and many things have not performed ths same ever since. Follow on generations of buffer cache hackers assumed the exising code was the reference and have been in frustrating state of coprophagia ever since. This commit 0) Brings back a (small) reserve allotment of buffer pages, and the kva to map them, to allow the cleaner and syncer to run even when under intense memory or kva pressure. 1) Fixes a lot of comments and variables to represent reality. 2) Simplifies and corrects how the buffer cache backs off down to the lowest level. 3) Corrects how the page daemons asks the buffer cache to back off, ensuring that uvmpd_scan is done to recover inactive pages in low memory situaitons 4) Adds a high water mark to the pool used to allocate struct buf's 5) Correct the cleaner and the sleep/wakeup cases in both low memory and low kva situations. (including accounting for the cleaner/syncer reserve) Tested by many, with very much helpful input from deraadt, miod, tobiasu, kettenis and others. ok kettenis@ deraadt@ jj@
Diffstat (limited to 'sys')
-rw-r--r--sys/kern/vfs_bio.c284
-rw-r--r--sys/uvm/uvm_pdaemon.c31
2 files changed, 188 insertions, 127 deletions
diff --git a/sys/kern/vfs_bio.c b/sys/kern/vfs_bio.c
index a41685a2e8a..93d3436ef60 100644
--- a/sys/kern/vfs_bio.c
+++ b/sys/kern/vfs_bio.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: vfs_bio.c,v 1.138 2012/10/16 02:30:54 beck Exp $ */
+/* $OpenBSD: vfs_bio.c,v 1.139 2012/11/07 17:50:48 beck Exp $ */
/* $NetBSD: vfs_bio.c,v 1.44 1996/06/11 11:15:36 pk Exp $ */
/*
@@ -69,6 +69,7 @@
#define BQ_CLEAN 1 /* LRU queue with clean buffers */
TAILQ_HEAD(bqueues, buf) bufqueues[BQUEUES];
+int nobuffers;
int needbuffer;
struct bio_ops bioops;
@@ -92,15 +93,27 @@ void bread_cluster_callback(struct buf *);
struct bcachestats bcstats; /* counters */
long lodirtypages; /* dirty page count low water mark */
long hidirtypages; /* dirty page count high water mark */
-long locleanpages; /* clean page count low water mark */
-long hicleanpages; /* clean page count high water mark */
-long backoffpages; /* backoff counter for page allocations */
-long buflowpages; /* bufpages low water mark */
-long bufhighpages; /* bufpages high water mark */
+long lopages; /* page recycling low water mark */
+long hipages; /* page recycling high water mark */
+long buflowpages; /* bufpages absolute low water mark */
+long bufhighpages; /* bufpages absolute high water mark */
long bufbackpages; /* number of pages we back off when asked to shrink */
vsize_t bufkvm;
+/*
+ * RESERVE_SLOTS of kva space, and the corresponding amount
+ * of buffer pages are reserved for the cleaner and syncer's
+ * exclusive use. Since we reserve kva slots to map the buffers
+ * along with the buffer space, this ensures the cleaner and
+ * syncer can always map and push out buffers if we get low
+ * on buffer pages or kva space in which to map them.
+ */
+#define RESERVE_SLOTS 4
+#define RESERVE_PAGES (RESERVE_SLOTS * MAXPHYS / PAGE_SIZE)
+#define BCACHE_MIN (RESERVE_PAGES * 2)
+#define UNCLEAN_PAGES (bcstats.numbufpages - bcstats.numcleanpages)
+
struct proc *cleanerproc;
int bd_req; /* Sleep point for cleaner daemon. */
@@ -156,11 +169,6 @@ buf_put(struct buf *bp)
LIST_REMOVE(bp, b_list);
bcstats.numbufs--;
- if (backoffpages) {
- backoffpages -= atop(bp->b_bufsize);
- if (backoffpages < 0)
- backoffpages = 0;
- }
if (buf_dealloc_mem(bp) != 0)
return;
@@ -177,24 +185,37 @@ bufinit(void)
struct bqueues *dp;
dmapages = uvm_pagecount(&dma_constraint);
+ /* take away a guess at how much of this the kernel will consume */
+ dmapages -= (atop(physmem) - atop(uvmexp.free));
/*
- * If MD code doesn't say otherwise, use 10% of kvm for mappings and
- * 10% of dmaable pages for cache pages.
+ * If MD code doesn't say otherwise, use up to 10% of DMA'able
+ * memory for buffers.
*/
if (bufcachepercent == 0)
bufcachepercent = 10;
+
+ /*
+ * XXX these values and their same use in kern_sysctl
+ * need to move into buf.h
+ */
+ KASSERT(bufcachepercent <= 90);
+ KASSERT(bufcachepercent >= 5);
if (bufpages == 0)
bufpages = dmapages * bufcachepercent / 100;
+ if (bufpages < BCACHE_MIN)
+ bufpages = BCACHE_MIN;
+ KASSERT(bufpages < dmapages);
bufhighpages = bufpages;
/*
- * set the base backoff level for the buffer cache to bufpages.
- * we will not allow uvm to steal back more than this number of
- * pages
+ * Set the base backoff level for the buffer cache. We will
+ * not allow uvm to steal back more than this number of pages.
*/
- buflowpages = dmapages * 10 / 100;
+ buflowpages = dmapages * 5 / 100;
+ if (buflowpages < BCACHE_MIN)
+ buflowpages = BCACHE_MIN;
/*
* set bufbackpages to 100 pages, or 10 percent of the low water mark
@@ -205,6 +226,10 @@ bufinit(void)
if (bufbackpages > 100)
bufbackpages = 100;
+ /*
+ * If the MD code does not say otherwise, reserve 10% of kva
+ * space for mapping buffers.
+ */
if (bufkvm == 0)
bufkvm = (VM_MAX_KERNEL_ADDRESS - VM_MIN_KERNEL_ADDRESS) / 10;
@@ -222,6 +247,8 @@ bufinit(void)
pool_init(&bufpool, sizeof(struct buf), 0, 0, 0, "bufpl", NULL);
pool_setipl(&bufpool, IPL_BIO);
+ pool_sethiwat(&bufpool, buflowpages / 4);
+
for (dp = bufqueues; dp < &bufqueues[BQUEUES]; dp++)
TAILQ_INIT(dp);
@@ -231,15 +258,21 @@ bufinit(void)
*/
buf_mem_init(bufkvm);
- hidirtypages = (bufpages / 4) * 3;
- lodirtypages = bufpages / 2;
+ /*
+ * Set the dirty page high water mark to be less than the low
+ * water mark for pages in the buffer cache. This ensures we
+ * can always back off by throwing away clean pages, and give
+ * ourselves a chance to write out the dirty pages eventually.
+ */
+ hidirtypages = (buflowpages / 4) * 3;
+ lodirtypages = buflowpages / 2;
/*
- * When we hit 95% of pages being clean, we bring them down to
- * 90% to have some slack.
+ * We are allowed to use up to the reserve. When we hit it,
+ * we free 10% of the cache size to allow room to recycle.
*/
- hicleanpages = bufpages - (bufpages / 20);
- locleanpages = bufpages - (bufpages / 10);
+ hipages = bufpages - RESERVE_PAGES;
+ lopages = hipages - (hipages / 10);
}
/*
@@ -248,33 +281,31 @@ bufinit(void)
void
bufadjust(int newbufpages)
{
- /*
- * XXX - note, bufkvm was allocated once, based on 10% of physmem
- * see above.
- */
struct buf *bp;
- int s;
+ int s, growing = 0;
+
+ if (newbufpages < buflowpages)
+ newbufpages = buflowpages;
s = splbio();
+ if (newbufpages >= bufpages)
+ growing = 1;
bufpages = newbufpages;
- hidirtypages = (bufpages / 4) * 3;
- lodirtypages = bufpages / 2;
-
/*
- * When we hit 95% of pages being clean, we bring them down to
- * 90% to have some slack.
+ * We are allowed to use up to the reserve. When we hit it,
+ * we free 10% of the cache size to allow room to recycle.
*/
- hicleanpages = bufpages - (bufpages / 20);
- locleanpages = bufpages - (bufpages / 10);
+ hipages = bufpages - RESERVE_PAGES;
+ lopages = hipages - (hipages / 10);
/*
- * If we we have more buffers allocated than bufpages,
- * free them up to get back down. this may possibly consume
- * all our clean pages...
+ * If we are shrinking the cache we are under some memory pressure.
+ * If we have more buffers allocated than our new low water mark,
+ * immediately free them.
*/
- while ((bp = TAILQ_FIRST(&bufqueues[BQ_CLEAN])) &&
- (bcstats.numbufpages > bufpages)) {
+ while (!growing && (bp = TAILQ_FIRST(&bufqueues[BQ_CLEAN])) &&
+ (bcstats.numbufpages > lopages)) {
bremfree(bp);
if (bp->b_vp) {
RB_REMOVE(buf_rb_bufs,
@@ -285,23 +316,13 @@ bufadjust(int newbufpages)
}
/*
- * Wake up cleaner if we're getting low on pages. We might
- * now have too much dirty, or have fallen below our low
- * water mark on clean pages so we need to free more stuff
- * up.
+ * Wake up the cleaner if we have lots of dirty pages,
+ * or if we are getting low on buffer cache kva.
*/
- if (bcstats.numdirtypages >= hidirtypages)
+ if (UNCLEAN_PAGES >= hidirtypages ||
+ bcstats.kvaslots_avail <= 2 * RESERVE_SLOTS)
wakeup(&bd_req);
- /*
- * if immediate action has not freed up enough goo for us
- * to proceed - we tsleep and wait for the cleaner above
- * to do it's work and get us reduced down to sanity.
- */
- while (bcstats.numbufpages > bufpages) {
- needbuffer++;
- tsleep(&needbuffer, PRIBIO, "needbuffer", 0);
- }
splx(s);
}
@@ -312,30 +333,33 @@ int
bufbackoff(struct uvm_constraint_range *range, long size)
{
/*
- * Back off the amount of buffer cache pages. Called by the page
- * daemon to consume buffer cache pages rather than swapping.
+ * Back off "size" buffer cache pages. Called by the page
+ * daemon to consume buffer cache pages rather than scanning.
*
- * On success, it frees N pages from the buffer cache, and sets
- * a flag so that the next N allocations from buf_get will recycle
- * a buffer rather than allocate a new one. It then returns 0 to the
- * caller.
+ * It returns 0 to the pagedaemon to indicate that it has
+ * succeeded in freeing enough pages. It returns -1 to
+ * indicate that it could not and the pagedaemon should take
+ * other measures.
*
- * on failure, it could free no pages from the buffer cache, does
- * nothing and returns -1 to the caller.
*/
- long d;
+ long pdelta, oldbufpages;
+
+ /*
+ * Back off by at least bufbackpages. If the page daemon gave us
+ * a larger size, back off by that much.
+ */
+ pdelta = (size > bufbackpages) ? size : bufbackpages;
if (bufpages <= buflowpages)
return(-1);
-
- if (bufpages - bufbackpages >= buflowpages)
- d = bufbackpages;
+ if (bufpages - pdelta < buflowpages)
+ pdelta = bufpages - buflowpages;
+ oldbufpages = bufpages;
+ bufadjust(bufpages - pdelta);
+ if (oldbufpages - bufpages < size)
+ return (-1); /* we did not free what we were asked */
else
- d = bufpages - buflowpages;
- backoffpages = bufbackpages;
- bufadjust(bufpages - d);
- backoffpages = 0;
- return(0);
+ return(0);
}
struct buf *
@@ -820,9 +844,16 @@ brelse(struct buf *bp)
CLR(bp->b_flags, (B_AGE | B_ASYNC | B_NOCACHE | B_DEFERRED));
buf_release(bp);
+ /* Wake up syncer and cleaner processes waiting for buffers. */
+ if (nobuffers) {
+ nobuffers = 0;
+ wakeup(&nobuffers);
+ }
+
/* Wake up any processes waiting for any buffer to become free. */
- if (needbuffer) {
- needbuffer--;
+ if (needbuffer && bcstats.numbufpages < hipages &&
+ bcstats.kvaslots_avail > RESERVE_SLOTS) {
+ needbuffer = 0;
wakeup(&needbuffer);
}
@@ -935,39 +966,43 @@ geteblk(int size)
struct buf *
buf_get(struct vnode *vp, daddr64_t blkno, size_t size)
{
- static int gcount = 0;
struct buf *bp;
int poolwait = size == 0 ? PR_NOWAIT : PR_WAITOK;
int npages;
int s;
- /*
- * if we were previously backed off, slowly climb back up
- * to the high water mark again.
- */
- if (backoffpages == 0 && bufpages < bufhighpages) {
- if (gcount == 0) {
- bufadjust(bufpages + bufbackpages);
- gcount += bufbackpages;
- } else
- gcount--;
- }
-
s = splbio();
if (size) {
/*
- * Wake up cleaner if we're getting low on pages.
+ * Wake up the cleaner if we have lots of dirty pages,
+ * or if we are getting low on buffer cache kva.
*/
- if (bcstats.numdirtypages >= hidirtypages)
+ if (UNCLEAN_PAGES >= hidirtypages ||
+ bcstats.kvaslots_avail <= 2 * RESERVE_SLOTS)
wakeup(&bd_req);
+ npages = atop(round_page(size));
+
+ /*
+ * If our allocation would take us over the
+ * high water mark, see if we can grow the
+ * cache.
+ */
+ if (bcstats.numbufpages + npages > hipages &&
+ bufpages < bufhighpages) {
+ int i = bufbackpages;
+ if (bufpages + i > bufhighpages)
+ i = bufhighpages - bufpages;
+ bufadjust(bufpages + i);
+ }
+
/*
- * If we're above the high water mark for clean pages,
+ * If we're still above the high water mark for pages,
* free down to the low water mark.
*/
- if (bcstats.numcleanpages > hicleanpages) {
- while (bcstats.numcleanpages > locleanpages) {
- bp = TAILQ_FIRST(&bufqueues[BQ_CLEAN]);
+ if (bcstats.numbufpages + npages > hipages) {
+ while ((bcstats.numbufpages > lopages) &&
+ (bp = TAILQ_FIRST(&bufqueues[BQ_CLEAN]))) {
bremfree(bp);
if (bp->b_vp) {
RB_REMOVE(buf_rb_bufs,
@@ -978,32 +1013,26 @@ buf_get(struct vnode *vp, daddr64_t blkno, size_t size)
}
}
- npages = atop(round_page(size));
-
/*
- * Free some buffers until we have enough space.
+ * If we get here, we tried to free the world down
+ * above, and couldn't get down - Wake the cleaner
+ * and wait for it to push some buffers out.
*/
- while ((bcstats.numbufpages + npages > bufpages)
- || backoffpages) {
- int freemax = 5;
- int i = freemax;
- while ((bp = TAILQ_FIRST(&bufqueues[BQ_CLEAN])) && i--) {
- bremfree(bp);
- if (bp->b_vp) {
- RB_REMOVE(buf_rb_bufs,
- &bp->b_vp->v_bufs_tree, bp);
- brelvp(bp);
- }
- buf_put(bp);
- }
- if (freemax == i &&
- (bcstats.numbufpages + npages > bufpages ||
- backoffpages)) {
- needbuffer++;
- tsleep(&needbuffer, PRIBIO, "needbuffer", 0);
- splx(s);
- return (NULL);
- }
+ if ((bcstats.numbufpages + npages > hipages ||
+ bcstats.kvaslots_avail <= RESERVE_SLOTS) &&
+ curproc != syncerproc && curproc != cleanerproc) {
+ wakeup(&bd_req);
+ needbuffer++;
+ tsleep(&needbuffer, PRIBIO, "needbuffer", 0);
+ splx(s);
+ return (NULL);
+ }
+ if (bcstats.numbufpages + npages > bufpages) {
+ /* cleaner or syncer */
+ nobuffers = 1;
+ tsleep(&nobuffers, PRIBIO, "nobuffers", 0);
+ splx(s);
+ return (NULL);
}
}
@@ -1068,21 +1097,35 @@ buf_daemon(struct proc *p)
{
struct timeval starttime, timediff;
struct buf *bp;
- int s;
+ int s, pushed;
cleanerproc = curproc;
+ pushed = 16;
s = splbio();
for (;;) {
- if (bcstats.numdirtypages < hidirtypages)
+ if (pushed >= 16 && (UNCLEAN_PAGES < hidirtypages &&
+ bcstats.kvaslots_avail > 2 * RESERVE_SLOTS)){
+ pushed = 0;
+ /*
+ * Wake up anyone who was waiting for buffers
+ * to be released.
+ */
+ if (needbuffer) {
+ needbuffer = 0;
+ wakeup(&needbuffer);
+ }
tsleep(&bd_req, PRIBIO - 7, "cleaner", 0);
+ }
getmicrouptime(&starttime);
while ((bp = TAILQ_FIRST(&bufqueues[BQ_DIRTY]))) {
struct timeval tv;
- if (bcstats.numdirtypages < lodirtypages)
+ if (UNCLEAN_PAGES < lodirtypages &&
+ bcstats.kvaslots_avail > 2 * RESERVE_SLOTS &&
+ pushed >= 16)
break;
bremfree(bp);
@@ -1111,6 +1154,7 @@ buf_daemon(struct proc *p)
}
bawrite(bp);
+ pushed++;
/* Never allow processing to run for more than 1 sec */
getmicrouptime(&tv);
@@ -1120,6 +1164,8 @@ buf_daemon(struct proc *p)
break;
}
+ if (bp == NULL)
+ pushed = 16; /* No dirty bufs - sleep */
}
}
diff --git a/sys/uvm/uvm_pdaemon.c b/sys/uvm/uvm_pdaemon.c
index 77b204ea52a..47f85afc638 100644
--- a/sys/uvm/uvm_pdaemon.c
+++ b/sys/uvm/uvm_pdaemon.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: uvm_pdaemon.c,v 1.59 2011/07/06 19:50:38 beck Exp $ */
+/* $OpenBSD: uvm_pdaemon.c,v 1.60 2012/11/07 17:50:49 beck Exp $ */
/* $NetBSD: uvm_pdaemon.c,v 1.23 2000/08/20 10:24:14 bjh21 Exp $ */
/*
@@ -247,13 +247,28 @@ uvm_pageout(void *arg)
if (pma != NULL ||
((uvmexp.free - BUFPAGES_DEFICIT) < uvmexp.freetarg) ||
((uvmexp.inactive + BUFPAGES_INACT) < uvmexp.inactarg)) {
- if (bufbackoff(&constraint,
- (pma ? pma->pm_size : -1)) == 0)
- work_done = 1;
- else {
- uvmpd_scan();
- work_done = 1; /* we hope... */
- }
+ u_int64_t free, size;
+ free = uvmexp.free - BUFPAGES_DEFICIT;
+ /* start with the size of what we are asking for */
+ size = pma ? pma->pm_size : 0;
+ /*
+ * If we below our target, add twice the amount
+ * we are below the target to what we will ask
+ * the buffer cache to give back. We then ask
+ * the buffer cache to give us free pages by
+ * backing off.
+ */
+ if (uvmexp.freetarg - free > 0)
+ size += (uvmexp.freetarg - free) * 2;
+ (void) bufbackoff(&constraint, size);
+ /*
+ * Now we scan to ensure we meet all our targets,
+ * we will not swap at this point, unless the buffer
+ * cache could not back off enough for us to meet
+ * our free page targets.
+ */
+ uvmpd_scan();
+ work_done = 1; /* XXX we hope... */
}
/*