summaryrefslogtreecommitdiff
path: root/sys/uvm
diff options
context:
space:
mode:
authorTed Unangst <tedu@cvs.openbsd.org>2013-05-30 15:18:00 +0000
committerTed Unangst <tedu@cvs.openbsd.org>2013-05-30 15:18:00 +0000
commit2aa7e70a5014d44f1bb185e26dbc14bb71c72d06 (patch)
tree5bd09a84498aea3be4e8eccfa3739f5ff8b0091c /sys/uvm
parent940bed6848d33e4b3337a56737835aeb0589f0de (diff)
remove simple_locks from uvm code. ok beck deraadt
Diffstat (limited to 'sys/uvm')
-rw-r--r--sys/uvm/uvm.h7
-rw-r--r--sys/uvm/uvm_amap.c15
-rw-r--r--sys/uvm/uvm_anon.c50
-rw-r--r--sys/uvm/uvm_anon.h4
-rw-r--r--sys/uvm/uvm_aobj.c43
-rw-r--r--sys/uvm/uvm_device.c10
-rw-r--r--sys/uvm/uvm_fault.c28
-rw-r--r--sys/uvm/uvm_km.c7
-rw-r--r--sys/uvm/uvm_map.c22
-rw-r--r--sys/uvm/uvm_map.h3
-rw-r--r--sys/uvm/uvm_mmap.c8
-rw-r--r--sys/uvm/uvm_object.c9
-rw-r--r--sys/uvm/uvm_object.h3
-rw-r--r--sys/uvm/uvm_page.c3
-rw-r--r--sys/uvm/uvm_page.h6
-rw-r--r--sys/uvm/uvm_pager.c43
-rw-r--r--sys/uvm/uvm_pdaemon.c67
-rw-r--r--sys/uvm/uvm_swap.c43
-rw-r--r--sys/uvm/uvm_vnode.c44
19 files changed, 21 insertions, 394 deletions
diff --git a/sys/uvm/uvm.h b/sys/uvm/uvm.h
index 14dac14dde2..5a68492a528 100644
--- a/sys/uvm/uvm.h
+++ b/sys/uvm/uvm.h
@@ -1,4 +1,4 @@
-/* $OpenBSD: uvm.h,v 1.48 2013/05/29 22:23:01 tedu Exp $ */
+/* $OpenBSD: uvm.h,v 1.49 2013/05/30 15:17:59 tedu Exp $ */
/* $NetBSD: uvm.h,v 1.24 2000/11/27 08:40:02 chs Exp $ */
/*
@@ -80,7 +80,6 @@ struct uvm {
struct pglist page_inactive_swp;/* pages inactive (reclaim or free) */
struct pglist page_inactive_obj;/* pages inactive (reclaim or free) */
/* Lock order: object lock, pageqlock, then fpageqlock. */
- simple_lock_data_t pageqlock; /* lock for active/inactive page q */
struct mutex fpageqlock; /* lock for free page q + pdaemon */
boolean_t page_init_done; /* TRUE if uvm_page_init() finished */
boolean_t page_idle_zero; /* TRUE if we should try to zero
@@ -98,14 +97,10 @@ struct uvm {
/* static kernel map entry pool */
vm_map_entry_t kentry_free; /* free page pool */
- simple_lock_data_t kentry_lock;
/* aio_done is locked by uvm.aiodoned_lock. */
TAILQ_HEAD(, buf) aio_done; /* done async i/o reqs */
- /* swap-related items */
- simple_lock_data_t swap_data_lock;
-
/* kernel object: to support anonymous pageable kernel memory */
struct uvm_object *kernel_object;
};
diff --git a/sys/uvm/uvm_amap.c b/sys/uvm/uvm_amap.c
index 73184c14ded..0433096ccbe 100644
--- a/sys/uvm/uvm_amap.c
+++ b/sys/uvm/uvm_amap.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: uvm_amap.c,v 1.47 2013/05/23 01:42:59 tedu Exp $ */
+/* $OpenBSD: uvm_amap.c,v 1.48 2013/05/30 15:17:59 tedu Exp $ */
/* $NetBSD: uvm_amap.c,v 1.27 2000/11/25 06:27:59 chs Exp $ */
/*
@@ -509,10 +509,7 @@ amap_wipeout(struct vm_amap *amap)
if (anon == NULL || anon->an_ref == 0)
panic("amap_wipeout: corrupt amap");
- simple_lock(&anon->an_lock); /* lock anon */
-
refs = --anon->an_ref;
- simple_unlock(&anon->an_lock);
if (refs == 0) {
/*
* we had the last reference to a vm_anon. free it.
@@ -633,9 +630,7 @@ amap_copy(struct vm_map *map, struct vm_map_entry *entry, int waitf,
srcamap->am_anon[entry->aref.ar_pageoff + lcv];
if (amap->am_anon[lcv] == NULL)
continue;
- simple_lock(&amap->am_anon[lcv]->an_lock);
amap->am_anon[lcv]->an_ref++;
- simple_unlock(&amap->am_anon[lcv]->an_lock);
amap->am_bckptr[lcv] = amap->am_nused;
amap->am_slots[amap->am_nused] = lcv;
amap->am_nused++;
@@ -715,7 +710,6 @@ ReStart:
slot = amap->am_slots[lcv];
anon = amap->am_anon[slot];
- simple_lock(&anon->an_lock);
pg = anon->an_page;
/*
@@ -765,10 +759,8 @@ ReStart:
* we can't ...
*/
if (nanon) {
- simple_lock(&nanon->an_lock);
uvm_anfree(nanon);
}
- simple_unlock(&anon->an_lock);
uvm_wait("cownowpage");
goto ReStart;
}
@@ -793,7 +785,6 @@ ReStart:
uvm_unlock_pageq();
}
- simple_unlock(&anon->an_lock);
/*
* done with this anon, next ...!
*/
@@ -999,9 +990,7 @@ amap_wiperange(struct vm_amap *amap, int slotoff, int slots)
/*
* drop anon reference count
*/
- simple_lock(&anon->an_lock);
refs = --anon->an_ref;
- simple_unlock(&anon->an_lock);
if (refs == 0) {
/*
* we just eliminated the last reference to an anon.
@@ -1054,11 +1043,9 @@ amap_swap_off(int startslot, int endslot)
slot = am->am_slots[i];
anon = am->am_anon[slot];
- simple_lock(&anon->an_lock);
swslot = anon->an_swslot;
if (swslot < startslot || endslot <= swslot) {
- simple_unlock(&anon->an_lock);
continue;
}
diff --git a/sys/uvm/uvm_anon.c b/sys/uvm/uvm_anon.c
index fb2b01d27f3..0c6e176f161 100644
--- a/sys/uvm/uvm_anon.c
+++ b/sys/uvm/uvm_anon.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: uvm_anon.c,v 1.35 2011/07/03 18:34:14 oga Exp $ */
+/* $OpenBSD: uvm_anon.c,v 1.36 2013/05/30 15:17:59 tedu Exp $ */
/* $NetBSD: uvm_anon.c,v 1.10 2000/11/25 06:27:59 chs Exp $ */
/*
@@ -70,11 +70,9 @@ uvm_analloc(void)
anon = pool_get(&uvm_anon_pool, PR_NOWAIT);
if (anon) {
- simple_lock_init(&anon->an_lock);
anon->an_ref = 1;
anon->an_page = NULL;
anon->an_swslot = 0;
- simple_lock(&anon->an_lock);
}
return(anon);
}
@@ -125,7 +123,6 @@ uvm_anfree(struct vm_anon *anon)
pg->loan_count--;
pg->uanon = NULL;
uvm_unlock_pageq();
- simple_unlock(&pg->uobject->vmobjlock);
} else {
/*
@@ -150,10 +147,8 @@ uvm_anfree(struct vm_anon *anon)
}
if (pg == NULL && anon->an_swslot != 0) {
/* this page is no longer only in swap. */
- simple_lock(&uvm.swap_data_lock);
KASSERT(uvmexp.swpgonly > 0);
uvmexp.swpgonly--;
- simple_unlock(&uvm.swap_data_lock);
}
/*
@@ -208,7 +203,6 @@ struct vm_page *
uvm_anon_lockloanpg(struct vm_anon *anon)
{
struct vm_page *pg;
- boolean_t locked = FALSE;
/*
* loop while we have a resident page that has a non-zero loan count.
@@ -221,44 +215,6 @@ uvm_anon_lockloanpg(struct vm_anon *anon)
while (((pg = anon->an_page) != NULL) && pg->loan_count != 0) {
- /*
- * quickly check to see if the page has an object before
- * bothering to lock the page queues. this may also produce
- * a false positive result, but that's ok because we do a real
- * check after that.
- *
- * XXX: quick check -- worth it? need volatile?
- */
-
- if (pg->uobject) {
-
- uvm_lock_pageq();
- if (pg->uobject) { /* the "real" check */
- locked =
- simple_lock_try(&pg->uobject->vmobjlock);
- } else {
- /* object disowned before we got PQ lock */
- locked = TRUE;
- }
- uvm_unlock_pageq();
-
- /*
- * if we didn't get a lock (try lock failed), then we
- * toggle our anon lock and try again
- */
-
- if (!locked) {
- simple_unlock(&anon->an_lock);
-
- /*
- * someone locking the object has a chance to
- * lock us right now
- */
-
- simple_lock(&anon->an_lock);
- continue;
- }
- }
/*
* if page is un-owned [i.e. the object dropped its ownership],
@@ -350,9 +306,5 @@ uvm_anon_pagein(struct vm_anon *anon)
* unlock the anon and we're done.
*/
- simple_unlock(&anon->an_lock);
- if (uobj) {
- simple_unlock(&uobj->vmobjlock);
- }
return FALSE;
}
diff --git a/sys/uvm/uvm_anon.h b/sys/uvm/uvm_anon.h
index a2f927f24e1..effb7f4657a 100644
--- a/sys/uvm/uvm_anon.h
+++ b/sys/uvm/uvm_anon.h
@@ -1,4 +1,4 @@
-/* $OpenBSD: uvm_anon.h,v 1.15 2010/06/14 10:05:37 thib Exp $ */
+/* $OpenBSD: uvm_anon.h,v 1.16 2013/05/30 15:17:59 tedu Exp $ */
/* $NetBSD: uvm_anon.h,v 1.13 2000/12/27 09:17:04 chs Exp $ */
/*
@@ -53,8 +53,6 @@ struct vm_anon {
* PG_BUSY]
*/
int an_swslot;
-
- simple_lock_data_t an_lock;
};
/*
diff --git a/sys/uvm/uvm_aobj.c b/sys/uvm/uvm_aobj.c
index e6e0f7f15b1..f668980602d 100644
--- a/sys/uvm/uvm_aobj.c
+++ b/sys/uvm/uvm_aobj.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: uvm_aobj.c,v 1.55 2012/04/12 14:59:26 ariane Exp $ */
+/* $OpenBSD: uvm_aobj.c,v 1.56 2013/05/30 15:17:59 tedu Exp $ */
/* $NetBSD: uvm_aobj.c,v 1.39 2001/02/18 21:19:08 chs Exp $ */
/*
@@ -382,8 +382,6 @@ static void
uao_free(struct uvm_aobj *aobj)
{
- simple_unlock(&aobj->u_obj.vmobjlock);
-
if (UAO_USES_SWHASH(aobj)) {
int i, hashbuckets = aobj->u_swhashmask + 1;
@@ -411,9 +409,7 @@ uao_free(struct uvm_aobj *aobj)
* this page is no longer
* only in swap.
*/
- simple_lock(&uvm.swap_data_lock);
uvmexp.swpgonly--;
- simple_unlock(&uvm.swap_data_lock);
}
next = LIST_NEXT(elt, list);
@@ -435,9 +431,7 @@ uao_free(struct uvm_aobj *aobj)
uvm_swap_free(slot, 1);
/* this page is no longer only in swap. */
- simple_lock(&uvm.swap_data_lock);
uvmexp.swpgonly--;
- simple_unlock(&uvm.swap_data_lock);
}
}
free(aobj->u_swslots, M_UVMAOBJ);
@@ -575,9 +569,7 @@ uao_init(void)
void
uao_reference(struct uvm_object *uobj)
{
- simple_lock(&uobj->vmobjlock);
uao_reference_locked(uobj);
- simple_unlock(&uobj->vmobjlock);
}
/*
@@ -612,7 +604,6 @@ uao_reference_locked(struct uvm_object *uobj)
void
uao_detach(struct uvm_object *uobj)
{
- simple_lock(&uobj->vmobjlock);
uao_detach_locked(uobj);
}
@@ -635,13 +626,11 @@ uao_detach_locked(struct uvm_object *uobj)
* detaching from kernel_object is a noop.
*/
if (UVM_OBJ_IS_KERN_OBJECT(uobj)) {
- simple_unlock(&uobj->vmobjlock);
return;
}
uobj->uo_refs--; /* drop ref! */
if (uobj->uo_refs) { /* still more refs? */
- simple_unlock(&uobj->vmobjlock);
return;
}
@@ -664,7 +653,6 @@ uao_detach_locked(struct uvm_object *uobj)
uvm_unlock_pageq();
UVM_UNLOCK_AND_WAIT(pg, &uobj->vmobjlock, 0,
"uao_det", 0);
- simple_lock(&uobj->vmobjlock);
uvm_lock_pageq();
continue;
}
@@ -747,7 +735,6 @@ uao_flush(struct uvm_object *uobj, voff_t start, voff_t stop, int flags)
atomic_setbits_int(&pp->pg_flags, PG_WANTED);
UVM_UNLOCK_AND_WAIT(pp, &uobj->vmobjlock, 0,
"uaoflsh", 0);
- simple_lock(&uobj->vmobjlock);
curoff -= PAGE_SIZE;
continue;
}
@@ -971,9 +958,7 @@ uao_get(struct uvm_object *uobj, voff_t offset, struct vm_page **pps,
/* out of RAM? */
if (ptmp == NULL) {
- simple_unlock(&uobj->vmobjlock);
uvm_wait("uao_getpage");
- simple_lock(&uobj->vmobjlock);
/* goto top of pps while loop */
continue;
}
@@ -996,7 +981,6 @@ uao_get(struct uvm_object *uobj, voff_t offset, struct vm_page **pps,
atomic_setbits_int(&ptmp->pg_flags, PG_WANTED);
UVM_UNLOCK_AND_WAIT(ptmp, &uobj->vmobjlock,
FALSE, "uao_get", 0);
- simple_lock(&uobj->vmobjlock);
continue; /* goto top of pps while loop */
}
@@ -1038,9 +1022,7 @@ uao_get(struct uvm_object *uobj, voff_t offset, struct vm_page **pps,
* page in the swapped-out page.
* unlock object for i/o, relock when done.
*/
- simple_unlock(&uobj->vmobjlock);
rv = uvm_swap_get(ptmp, swslot, PGO_SYNCIO);
- simple_lock(&uobj->vmobjlock);
/*
* I/O done. check for errors.
@@ -1067,7 +1049,6 @@ uao_get(struct uvm_object *uobj, voff_t offset, struct vm_page **pps,
uvm_pagefree(ptmp);
uvm_unlock_pageq();
- simple_unlock(&uobj->vmobjlock);
return (rv);
}
}
@@ -1094,7 +1075,6 @@ uao_get(struct uvm_object *uobj, voff_t offset, struct vm_page **pps,
* finally, unlock object and return.
*/
- simple_unlock(&uobj->vmobjlock);
return(VM_PAGER_OK);
}
@@ -1132,7 +1112,6 @@ uao_swap_off(int startslot, int endslot)
* walk the list of all aobjs.
*/
-restart:
mtx_enter(&uao_list_lock);
for (aobj = LIST_FIRST(&uao_list);
@@ -1141,21 +1120,6 @@ restart:
boolean_t rv;
/*
- * try to get the object lock,
- * start all over if we fail.
- * most of the time we'll get the aobj lock,
- * so this should be a rare case.
- */
- if (!simple_lock_try(&aobj->u_obj.vmobjlock)) {
- mtx_leave(&uao_list_lock);
- if (prevaobj) {
- uao_detach_locked(&prevaobj->u_obj);
- prevaobj = NULL;
- }
- goto restart;
- }
-
- /*
* add a ref to the aobj so it doesn't disappear
* while we're working.
*/
@@ -1302,11 +1266,6 @@ uao_pagein_page(struct uvm_aobj *aobj, int pageidx)
&pg, &npages, 0, VM_PROT_READ|VM_PROT_WRITE, 0, 0);
/* unlocked: aobj */
- /*
- * relock and finish up.
- */
- simple_lock(&aobj->u_obj.vmobjlock);
-
switch (rv) {
case VM_PAGER_OK:
break;
diff --git a/sys/uvm/uvm_device.c b/sys/uvm/uvm_device.c
index 2cb3aa00c3f..a32bc5cefca 100644
--- a/sys/uvm/uvm_device.c
+++ b/sys/uvm/uvm_device.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: uvm_device.c,v 1.40 2011/07/03 18:34:14 oga Exp $ */
+/* $OpenBSD: uvm_device.c,v 1.41 2013/05/30 15:17:59 tedu Exp $ */
/* $NetBSD: uvm_device.c,v 1.30 2000/11/25 06:27:59 chs Exp $ */
/*
@@ -180,9 +180,7 @@ udv_attach(void *arg, vm_prot_t accessprot, voff_t off, vsize_t size)
* bump reference count, unhold, return.
*/
- simple_lock(&lcv->u_obj.vmobjlock);
lcv->u_obj.uo_refs++;
- simple_unlock(&lcv->u_obj.vmobjlock);
mtx_enter(&udv_lock);
if (lcv->u_flags & UVM_DEVICE_WANTED)
@@ -251,9 +249,7 @@ static void
udv_reference(struct uvm_object *uobj)
{
- simple_lock(&uobj->vmobjlock);
uobj->uo_refs++;
- simple_unlock(&uobj->vmobjlock);
}
/*
@@ -273,10 +269,8 @@ udv_detach(struct uvm_object *uobj)
* loop until done
*/
again:
- simple_lock(&uobj->vmobjlock);
if (uobj->uo_refs > 1) {
uobj->uo_refs--;
- simple_unlock(&uobj->vmobjlock);
return;
}
KASSERT(uobj->uo_npages == 0 && RB_EMPTY(&uobj->memt));
@@ -292,7 +286,6 @@ again:
* lock interleaving. -- this is ok in this case since the
* locks are both IPL_NONE
*/
- simple_unlock(&uobj->vmobjlock);
msleep(udv, &udv_lock, PVM | PNORELOCK, "udv_detach", 0);
goto again;
}
@@ -305,7 +298,6 @@ again:
if (udv->u_flags & UVM_DEVICE_WANTED)
wakeup(udv);
mtx_leave(&udv_lock);
- simple_unlock(&uobj->vmobjlock);
free(udv, M_TEMP);
}
diff --git a/sys/uvm/uvm_fault.c b/sys/uvm/uvm_fault.c
index ed0c2cf4202..b7d53d31197 100644
--- a/sys/uvm/uvm_fault.c
+++ b/sys/uvm/uvm_fault.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: uvm_fault.c,v 1.65 2012/04/12 11:55:43 ariane Exp $ */
+/* $OpenBSD: uvm_fault.c,v 1.66 2013/05/30 15:17:59 tedu Exp $ */
/* $NetBSD: uvm_fault.c,v 1.51 2000/08/06 00:22:53 thorpej Exp $ */
/*
@@ -199,7 +199,6 @@ uvmfault_anonflush(struct vm_anon **anons, int n)
for (lcv = 0 ; lcv < n ; lcv++) {
if (anons[lcv] == NULL)
continue;
- simple_lock(&anons[lcv]->an_lock);
pg = anons[lcv]->an_page;
if (pg && (pg->pg_flags & PG_BUSY) == 0 && pg->loan_count == 0) {
uvm_lock_pageq();
@@ -213,7 +212,6 @@ uvmfault_anonflush(struct vm_anon **anons, int n)
}
uvm_unlock_pageq();
}
- simple_unlock(&anons[lcv]->an_lock);
}
}
@@ -405,8 +403,6 @@ uvmfault_anonget(struct uvm_faultinfo *ufi, struct vm_amap *amap,
*/
locked = uvmfault_relock(ufi);
- if (locked || we_own)
- simple_lock(&anon->an_lock);
/*
* if we own the page (i.e. we set PG_BUSY), then we need
@@ -437,7 +433,6 @@ uvmfault_anonget(struct uvm_faultinfo *ufi, struct vm_amap *amap,
*/
if (pg->pg_flags & PG_RELEASED) {
pmap_page_protect(pg, VM_PROT_NONE);
- simple_unlock(&anon->an_lock);
uvm_anfree(anon); /* frees page for us */
if (locked)
uvmfault_unlockall(ufi, amap, NULL,
@@ -473,8 +468,6 @@ uvmfault_anonget(struct uvm_faultinfo *ufi, struct vm_amap *amap,
if (locked)
uvmfault_unlockall(ufi, amap, NULL,
anon);
- else
- simple_unlock(&anon->an_lock);
return (VM_PAGER_ERROR);
}
@@ -486,8 +479,6 @@ uvmfault_anonget(struct uvm_faultinfo *ufi, struct vm_amap *amap,
uvm_lock_pageq();
uvm_pageactivate(pg);
uvm_unlock_pageq();
- if (!locked)
- simple_unlock(&anon->an_lock);
}
/*
@@ -768,10 +759,8 @@ ReFault:
/* flush object? */
if (uobj) {
uoff = (startva - ufi.entry->start) + ufi.entry->offset;
- simple_lock(&uobj->vmobjlock);
(void) uobj->pgops->pgo_flush(uobj, uoff, uoff +
(nback << PAGE_SHIFT), PGO_DEACTIVATE);
- simple_unlock(&uobj->vmobjlock);
}
/* now forget about the backpages */
@@ -822,7 +811,6 @@ ReFault:
continue;
}
anon = anons[lcv];
- simple_lock(&anon->an_lock);
/* ignore loaned pages */
if (anon->an_page && anon->an_page->loan_count == 0 &&
(anon->an_page->pg_flags & (PG_RELEASED|PG_BUSY)) == 0) {
@@ -844,7 +832,6 @@ ReFault:
PMAP_CANFAIL |
(VM_MAPENT_ISWIRED(ufi.entry) ? PMAP_WIRED : 0));
}
- simple_unlock(&anon->an_lock);
pmap_update(ufi.orig_map->pmap);
}
@@ -868,8 +855,6 @@ ReFault:
*/
if (uobj && shadowed == FALSE && uobj->pgops->pgo_fault != NULL) {
- simple_lock(&uobj->vmobjlock);
-
/* locked: maps(read), amap (if there), uobj */
result = uobj->pgops->pgo_fault(&ufi, startva, pages, npages,
centeridx, fault_type, access_type,
@@ -896,8 +881,6 @@ ReFault:
*/
if (uobj && shadowed == FALSE) {
- simple_lock(&uobj->vmobjlock);
-
/* locked (!shadowed): maps(read), amap (if there), uobj */
/*
* the following call to pgo_get does _not_ change locking state
@@ -1018,7 +1001,6 @@ ReFault:
*/
anon = anons[centeridx];
- simple_lock(&anon->an_lock);
/* locked: maps(read), amap, anon */
@@ -1130,7 +1112,6 @@ ReFault:
uvm_pageactivate(pg);
uvm_unlock_pageq();
if (uobj) {
- simple_unlock(&uobj->vmobjlock);
uobj = NULL;
}
@@ -1353,7 +1334,6 @@ Case2:
*/
locked = uvmfault_relock(&ufi);
- simple_lock(&uobj->vmobjlock);
/* locked(locked): maps(read), amap(if !null), uobj, uobjpage */
/* locked(!locked): uobj, uobjpage */
@@ -1387,7 +1367,6 @@ Case2:
atomic_clearbits_int(&uobjpage->pg_flags,
PG_BUSY|PG_WANTED);
UVM_PAGE_OWN(uobjpage, NULL);
- simple_unlock(&uobj->vmobjlock);
goto ReFault;
}
@@ -1599,7 +1578,6 @@ Case2:
uvm_lock_pageq();
uvm_pageactivate(uobjpage);
uvm_unlock_pageq();
- simple_unlock(&uobj->vmobjlock);
uobj = NULL;
} else {
uvmexp.flt_przero++;
@@ -1830,10 +1808,6 @@ uvmfault_unlockall(struct uvm_faultinfo *ufi, struct vm_amap *amap,
struct uvm_object *uobj, struct vm_anon *anon)
{
- if (anon)
- simple_unlock(&anon->an_lock);
- if (uobj)
- simple_unlock(&uobj->vmobjlock);
uvmfault_unlockmaps(ufi, FALSE);
}
diff --git a/sys/uvm/uvm_km.c b/sys/uvm/uvm_km.c
index 3131053ecc9..2c8af0aa25d 100644
--- a/sys/uvm/uvm_km.c
+++ b/sys/uvm/uvm_km.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: uvm_km.c,v 1.108 2012/11/10 11:18:45 kettenis Exp $ */
+/* $OpenBSD: uvm_km.c,v 1.109 2013/05/30 15:17:59 tedu Exp $ */
/* $NetBSD: uvm_km.c,v 1.42 2001/01/14 02:10:01 thorpej Exp $ */
/*
@@ -282,7 +282,6 @@ uvm_km_pgremove(struct uvm_object *uobj, vaddr_t start, vaddr_t end)
atomic_setbits_int(&pp->pg_flags, PG_WANTED);
UVM_UNLOCK_AND_WAIT(pp, &uobj->vmobjlock, 0,
"km_pgrm", 0);
- simple_lock(&uobj->vmobjlock);
curoff -= PAGE_SIZE; /* loop back to us */
continue;
}
@@ -295,9 +294,7 @@ uvm_km_pgremove(struct uvm_object *uobj, vaddr_t start, vaddr_t end)
uvm_pagefree(pp);
uvm_unlock_pageq();
} else if (slot != 0) {
- simple_lock(&uvm.swap_data_lock);
uvmexp.swpgonly--;
- simple_unlock(&uvm.swap_data_lock);
}
}
}
@@ -522,14 +519,12 @@ uvm_km_alloc1(struct vm_map *map, vsize_t size, vsize_t align, boolean_t zeroit)
loopva = kva;
while (size) {
- simple_lock(&uvm.kernel_object->vmobjlock);
/* allocate ram */
pg = uvm_pagealloc(uvm.kernel_object, offset, NULL, 0);
if (pg) {
atomic_clearbits_int(&pg->pg_flags, PG_BUSY);
UVM_PAGE_OWN(pg, NULL);
}
- simple_unlock(&uvm.kernel_object->vmobjlock);
if (__predict_false(pg == NULL)) {
if (curproc == uvm.pagedaemon_proc) {
/*
diff --git a/sys/uvm/uvm_map.c b/sys/uvm/uvm_map.c
index e5496f4e0ee..af0aa631469 100644
--- a/sys/uvm/uvm_map.c
+++ b/sys/uvm/uvm_map.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: uvm_map.c,v 1.161 2013/04/17 23:22:42 tedu Exp $ */
+/* $OpenBSD: uvm_map.c,v 1.162 2013/05/30 15:17:59 tedu Exp $ */
/* $NetBSD: uvm_map.c,v 1.86 2000/11/27 08:40:03 chs Exp $ */
/*
@@ -491,9 +491,7 @@ uvm_mapent_addr_remove(struct vm_map *map, struct vm_map_entry *entry)
*/
#define uvm_map_reference(_map) \
do { \
- simple_lock(&map->ref_lock); \
map->ref_count++; \
- simple_unlock(&map->ref_lock); \
} while (0)
/*
@@ -1521,7 +1519,6 @@ uvm_mapent_alloc(struct vm_map *map, int flags)
if (map->flags & VM_MAP_INTRSAFE || cold) {
s = splvm();
- simple_lock(&uvm.kentry_lock);
me = uvm.kentry_free;
if (me == NULL) {
ne = km_alloc(PAGE_SIZE, &kv_page, &kp_dirty,
@@ -1542,7 +1539,6 @@ uvm_mapent_alloc(struct vm_map *map, int flags)
}
uvm.kentry_free = RB_LEFT(me, daddrs.addr_entry);
uvmexp.kmapent++;
- simple_unlock(&uvm.kentry_lock);
splx(s);
me->flags = UVM_MAP_STATIC;
} else if (map == kernel_map) {
@@ -1581,11 +1577,9 @@ uvm_mapent_free(struct vm_map_entry *me)
if (me->flags & UVM_MAP_STATIC) {
s = splvm();
- simple_lock(&uvm.kentry_lock);
RB_LEFT(me, daddrs.addr_entry) = uvm.kentry_free;
uvm.kentry_free = me;
uvmexp.kmapent--;
- simple_unlock(&uvm.kentry_lock);
splx(s);
} else if (me->flags & UVM_MAP_KMEM) {
splassert(IPL_NONE);
@@ -2295,7 +2289,6 @@ uvm_map_setup(struct vm_map *map, vaddr_t min, vaddr_t max, int flags)
map->flags = flags;
map->timestamp = 0;
rw_init(&map->lock, "vmmaplk");
- simple_lock_init(&map->ref_lock);
/*
* Configure the allocators.
@@ -2680,7 +2673,6 @@ uvm_map_init(void)
* now set up static pool of kernel map entries ...
*/
- simple_lock_init(&uvm.kentry_lock);
uvm.kentry_free = NULL;
for (lcv = 0 ; lcv < MAX_KMAPENT ; lcv++) {
RB_LEFT(&kernel_map_entry[lcv], daddrs.addr_entry) =
@@ -3788,9 +3780,7 @@ uvm_map_deallocate(vm_map_t map)
int c;
struct uvm_map_deadq dead;
- simple_lock(&map->ref_lock);
c = --map->ref_count;
- simple_unlock(&map->ref_lock);
if (c > 0) {
return;
}
@@ -4169,11 +4159,8 @@ uvm_map_clean(struct vm_map *map, vaddr_t start, vaddr_t end, int flags)
if (anon == NULL)
continue;
- simple_lock(&anon->an_lock); /* XXX */
-
pg = anon->an_page;
if (pg == NULL) {
- simple_unlock(&anon->an_lock);
continue;
}
@@ -4191,7 +4178,6 @@ deactivate_it:
/* skip the page if it's loaned or wired */
if (pg->loan_count != 0 ||
pg->wire_count != 0) {
- simple_unlock(&anon->an_lock);
break;
}
@@ -4205,7 +4191,6 @@ deactivate_it:
if ((pg->pg_flags & PQ_ANON) == 0) {
KASSERT(pg->uobject == NULL);
uvm_unlock_pageq();
- simple_unlock(&anon->an_lock);
break;
}
KASSERT(pg->uanon == anon);
@@ -4217,7 +4202,6 @@ deactivate_it:
uvm_pagedeactivate(pg);
uvm_unlock_pageq();
- simple_unlock(&anon->an_lock);
break;
case PGO_FREE:
@@ -4231,13 +4215,11 @@ deactivate_it:
/* XXX skip the page if it's wired */
if (pg->wire_count != 0) {
- simple_unlock(&anon->an_lock);
break;
}
amap_unadd(&entry->aref,
cp_start - entry->start);
refs = --anon->an_ref;
- simple_unlock(&anon->an_lock);
if (refs == 0)
uvm_anfree(anon);
break;
@@ -4262,11 +4244,9 @@ flush_object:
((flags & PGO_FREE) == 0 ||
((entry->max_protection & VM_PROT_WRITE) != 0 &&
(entry->etype & UVM_ET_COPYONWRITE) == 0))) {
- simple_lock(&uobj->vmobjlock);
rv = uobj->pgops->pgo_flush(uobj,
cp_start - entry->start + entry->offset,
cp_end - entry->start + entry->offset, flags);
- simple_unlock(&uobj->vmobjlock);
if (rv == FALSE)
error = EFAULT;
diff --git a/sys/uvm/uvm_map.h b/sys/uvm/uvm_map.h
index 9cf6827b609..236a533c46c 100644
--- a/sys/uvm/uvm_map.h
+++ b/sys/uvm/uvm_map.h
@@ -1,4 +1,4 @@
-/* $OpenBSD: uvm_map.h,v 1.48 2012/04/11 11:23:22 ariane Exp $ */
+/* $OpenBSD: uvm_map.h,v 1.49 2013/05/30 15:17:59 tedu Exp $ */
/* $NetBSD: uvm_map.h,v 1.24 2001/02/18 21:19:08 chs Exp $ */
/*
@@ -297,7 +297,6 @@ struct vm_map {
vsize_t size; /* virtual size */
int ref_count; /* Reference count */
- simple_lock_data_t ref_lock; /* Lock for ref_count field */
int flags; /* flags */
unsigned int timestamp; /* Version number */
diff --git a/sys/uvm/uvm_mmap.c b/sys/uvm/uvm_mmap.c
index 25c1306b87e..8ad58fb4501 100644
--- a/sys/uvm/uvm_mmap.c
+++ b/sys/uvm/uvm_mmap.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: uvm_mmap.c,v 1.91 2012/07/21 06:46:58 matthew Exp $ */
+/* $OpenBSD: uvm_mmap.c,v 1.92 2013/05/30 15:17:59 tedu Exp $ */
/* $NetBSD: uvm_mmap.c,v 1.49 2001/02/18 21:19:08 chs Exp $ */
/*
@@ -258,9 +258,6 @@ sys_mincore(struct proc *p, void *v, register_t *retval)
amap = entry->aref.ar_amap; /* top layer */
uobj = entry->object.uvm_obj; /* bottom layer */
- if (uobj != NULL)
- simple_lock(&uobj->vmobjlock);
-
for (/* nothing */; start < lim; start += PAGE_SIZE, pgi++) {
*pgi = 0;
if (amap != NULL) {
@@ -290,9 +287,6 @@ sys_mincore(struct proc *p, void *v, register_t *retval)
}
}
}
-
- if (uobj != NULL)
- simple_unlock(&uobj->vmobjlock);
}
out:
diff --git a/sys/uvm/uvm_object.c b/sys/uvm/uvm_object.c
index ffd76a259e5..e6844f08c52 100644
--- a/sys/uvm/uvm_object.c
+++ b/sys/uvm/uvm_object.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: uvm_object.c,v 1.6 2010/05/01 13:13:10 oga Exp $ */
+/* $OpenBSD: uvm_object.c,v 1.7 2013/05/30 15:17:59 tedu Exp $ */
/*
* Copyright (c) 2006 The NetBSD Foundation, Inc.
@@ -73,7 +73,6 @@ uvm_objwire(struct uvm_object *uobj, off_t start, off_t end,
left = (end - start) >> PAGE_SHIFT;
- simple_lock(&uobj->vmobjlock);
while (left) {
npages = MIN(FETCH_PAGECOUNT, left);
@@ -87,7 +86,6 @@ uvm_objwire(struct uvm_object *uobj, off_t start, off_t end,
if (error)
goto error;
- simple_lock(&uobj->vmobjlock);
for (i = 0; i < npages; i++) {
KASSERT(pgs[i] != NULL);
@@ -101,9 +99,7 @@ uvm_objwire(struct uvm_object *uobj, off_t start, off_t end,
while (pgs[i]->loan_count) {
pg = uvm_loanbreak(pgs[i]);
if (!pg) {
- simple_unlock(&uobj->vmobjlock);
uvm_wait("uobjwirepg");
- simple_lock(&uobj->vmobjlock);
continue;
}
}
@@ -133,7 +129,6 @@ uvm_objwire(struct uvm_object *uobj, off_t start, off_t end,
left -= npages;
offset += npages << PAGE_SHIFT;
}
- simple_unlock(&uobj->vmobjlock);
return 0;
@@ -156,7 +151,6 @@ uvm_objunwire(struct uvm_object *uobj, off_t start, off_t end)
struct vm_page *pg;
off_t offset;
- simple_lock(&uobj->vmobjlock);
uvm_lock_pageq();
for (offset = start; offset < end; offset += PAGE_SIZE) {
pg = uvm_pagelookup(uobj, offset);
@@ -167,6 +161,5 @@ uvm_objunwire(struct uvm_object *uobj, off_t start, off_t end)
uvm_pageunwire(pg);
}
uvm_unlock_pageq();
- simple_unlock(&uobj->vmobjlock);
}
#endif /* !SMALL_KERNEL */
diff --git a/sys/uvm/uvm_object.h b/sys/uvm/uvm_object.h
index 65ae4445f81..83085701bdb 100644
--- a/sys/uvm/uvm_object.h
+++ b/sys/uvm/uvm_object.h
@@ -1,4 +1,4 @@
-/* $OpenBSD: uvm_object.h,v 1.18 2011/07/07 14:25:15 sobrado Exp $ */
+/* $OpenBSD: uvm_object.h,v 1.19 2013/05/30 15:17:59 tedu Exp $ */
/* $NetBSD: uvm_object.h,v 1.11 2001/03/09 01:02:12 chs Exp $ */
/*
@@ -47,7 +47,6 @@
*/
struct uvm_object {
- simple_lock_data_t vmobjlock; /* lock on memt */
struct uvm_pagerops *pgops; /* pager ops */
RB_HEAD(uvm_objtree, vm_page) memt; /* pages in object */
int uo_npages; /* # of pages in memt */
diff --git a/sys/uvm/uvm_page.c b/sys/uvm/uvm_page.c
index 29a7fe92186..cdce265a4d0 100644
--- a/sys/uvm/uvm_page.c
+++ b/sys/uvm/uvm_page.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: uvm_page.c,v 1.123 2013/03/27 02:02:23 tedu Exp $ */
+/* $OpenBSD: uvm_page.c,v 1.124 2013/05/30 15:17:59 tedu Exp $ */
/* $NetBSD: uvm_page.c,v 1.44 2000/11/27 08:40:04 chs Exp $ */
/*
@@ -203,7 +203,6 @@ uvm_page_init(vaddr_t *kvm_startp, vaddr_t *kvm_endp)
TAILQ_INIT(&uvm.page_active);
TAILQ_INIT(&uvm.page_inactive_swp);
TAILQ_INIT(&uvm.page_inactive_obj);
- simple_lock_init(&uvm.pageqlock);
mtx_init(&uvm.fpageqlock, IPL_VM);
uvm_pmr_init();
diff --git a/sys/uvm/uvm_page.h b/sys/uvm/uvm_page.h
index 9ab6f1481c7..428f82e3a5f 100644
--- a/sys/uvm/uvm_page.h
+++ b/sys/uvm/uvm_page.h
@@ -1,4 +1,4 @@
-/* $OpenBSD: uvm_page.h,v 1.48 2011/05/30 22:25:24 oga Exp $ */
+/* $OpenBSD: uvm_page.h,v 1.49 2013/05/30 15:17:59 tedu Exp $ */
/* $NetBSD: uvm_page.h,v 1.19 2000/12/28 08:24:55 chs Exp $ */
/*
@@ -299,8 +299,8 @@ int vm_physseg_find(paddr_t, int *);
* macros
*/
-#define uvm_lock_pageq() simple_lock(&uvm.pageqlock)
-#define uvm_unlock_pageq() simple_unlock(&uvm.pageqlock)
+#define uvm_lock_pageq() /* lock */
+#define uvm_unlock_pageq() /* unlock */
#define uvm_lock_fpageq() mtx_enter(&uvm.fpageqlock);
#define uvm_unlock_fpageq() mtx_leave(&uvm.fpageqlock);
diff --git a/sys/uvm/uvm_pager.c b/sys/uvm/uvm_pager.c
index 389ab19725c..84ef7cf5e88 100644
--- a/sys/uvm/uvm_pager.c
+++ b/sys/uvm/uvm_pager.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: uvm_pager.c,v 1.60 2011/07/03 18:34:14 oga Exp $ */
+/* $OpenBSD: uvm_pager.c,v 1.61 2013/05/30 15:17:59 tedu Exp $ */
/* $NetBSD: uvm_pager.c,v 1.36 2000/11/27 18:26:41 chs Exp $ */
/*
@@ -567,9 +567,6 @@ ReTry:
* drop cluster and relock object (only if I/O is
* not pending)
*/
- if (uobj)
- /* required for dropcluster */
- simple_lock(&uobj->vmobjlock);
if (*npages > 1 || pg == NULL)
uvm_pager_dropcluster(uobj, pg, ppsp, npages,
PGO_PDFREECLUST);
@@ -586,9 +583,6 @@ ReTry:
*/
if (*npages > 1 || pg == NULL) {
- if (uobj) {
- simple_lock(&uobj->vmobjlock);
- }
uvm_pager_dropcluster(uobj, pg, ppsp, npages, PGO_REALLOCSWAP);
/*
@@ -602,15 +596,11 @@ ReTry:
/* XXX daddr64_t -> int */
int nswblk = (result == VM_PAGER_AGAIN) ? swblk : 0;
if (pg->pg_flags & PQ_ANON) {
- simple_lock(&pg->uanon->an_lock);
pg->uanon->an_swslot = nswblk;
- simple_unlock(&pg->uanon->an_lock);
} else {
- simple_lock(&pg->uobject->vmobjlock);
uao_set_swslot(pg->uobject,
pg->offset >> PAGE_SHIFT,
nswblk);
- simple_unlock(&pg->uobject->vmobjlock);
}
}
if (result == VM_PAGER_AGAIN) {
@@ -653,8 +643,6 @@ ReTry:
* to worry about.
*/
- if (uobj && (flags & PGO_PDFREECLUST) != 0)
- simple_lock(&uobj->vmobjlock);
return(result);
}
@@ -704,12 +692,10 @@ uvm_pager_dropcluster(struct uvm_object *uobj, struct vm_page *pg,
*/
if (!uobj) {
if (ppsp[lcv]->pg_flags & PQ_ANON) {
- simple_lock(&ppsp[lcv]->uanon->an_lock);
if (flags & PGO_REALLOCSWAP)
/* zap swap block */
ppsp[lcv]->uanon->an_swslot = 0;
} else {
- simple_lock(&ppsp[lcv]->uobject->vmobjlock);
if (flags & PGO_REALLOCSWAP)
uao_set_swslot(ppsp[lcv]->uobject,
ppsp[lcv]->offset >> PAGE_SHIFT, 0);
@@ -730,7 +716,6 @@ uvm_pager_dropcluster(struct uvm_object *uobj, struct vm_page *pg,
PG_BUSY);
UVM_PAGE_OWN(ppsp[lcv], NULL);
- simple_unlock(&ppsp[lcv]->uanon->an_lock);
/* kills anon and frees pg */
uvm_anfree(ppsp[lcv]->uanon);
@@ -754,14 +739,6 @@ uvm_pager_dropcluster(struct uvm_object *uobj, struct vm_page *pg,
pmap_clear_modify(ppsp[lcv]);
atomic_setbits_int(&ppsp[lcv]->pg_flags, PG_CLEAN);
}
-
- /* if anonymous cluster, unlock object and move on */
- if (!uobj) {
- if (ppsp[lcv]->pg_flags & PQ_ANON)
- simple_unlock(&ppsp[lcv]->uanon->an_lock);
- else
- simple_unlock(&ppsp[lcv]->uobject->vmobjlock);
- }
}
}
@@ -826,17 +803,9 @@ uvm_aio_aiodone(struct buf *bp)
swap = (pg->pg_flags & PQ_SWAPBACKED) != 0;
if (!swap) {
uobj = pg->uobject;
- simple_lock(&uobj->vmobjlock);
}
}
KASSERT(swap || pg->uobject == uobj);
- if (swap) {
- if (pg->pg_flags & PQ_ANON) {
- simple_lock(&pg->uanon->an_lock);
- } else {
- simple_lock(&pg->uobject->vmobjlock);
- }
- }
/*
* if this is a read and we got an error, mark the pages
@@ -860,18 +829,8 @@ uvm_aio_aiodone(struct buf *bp)
atomic_setbits_int(&pgs[i]->pg_flags, PG_CLEAN);
atomic_clearbits_int(&pgs[i]->pg_flags, PG_FAKE);
}
- if (swap) {
- if (pg->pg_flags & PQ_ANON) {
- simple_unlock(&pg->uanon->an_lock);
- } else {
- simple_unlock(&pg->uobject->vmobjlock);
- }
- }
}
uvm_page_unbusy(pgs, npages);
- if (!swap) {
- simple_unlock(&uobj->vmobjlock);
- }
#ifdef UVM_SWAP_ENCRYPT
freed:
diff --git a/sys/uvm/uvm_pdaemon.c b/sys/uvm/uvm_pdaemon.c
index d1ecf3e7284..aadaa598b72 100644
--- a/sys/uvm/uvm_pdaemon.c
+++ b/sys/uvm/uvm_pdaemon.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: uvm_pdaemon.c,v 1.62 2013/02/07 17:29:31 beck Exp $ */
+/* $OpenBSD: uvm_pdaemon.c,v 1.63 2013/05/30 15:17:59 tedu Exp $ */
/* $NetBSD: uvm_pdaemon.c,v 1.23 2000/08/20 10:24:14 bjh21 Exp $ */
/*
@@ -465,10 +465,6 @@ uvmpd_scan_inactive(struct pglist *pglst)
if ((p->pg_flags & PQ_ANON) || p->uobject == NULL) {
anon = p->uanon;
KASSERT(anon != NULL);
- if (!simple_lock_try(&anon->an_lock)) {
- /* lock failed, skip this page */
- continue;
- }
/*
* if the page is ownerless, claim it in the
@@ -483,7 +479,6 @@ uvmpd_scan_inactive(struct pglist *pglst)
/* anon now owns it */
}
if (p->pg_flags & PG_BUSY) {
- simple_unlock(&anon->an_lock);
uvmexp.pdbusy++;
/* someone else owns page, skip it */
continue;
@@ -492,12 +487,7 @@ uvmpd_scan_inactive(struct pglist *pglst)
} else {
uobj = p->uobject;
KASSERT(uobj != NULL);
- if (!simple_lock_try(&uobj->vmobjlock)) {
- /* lock failed, skip this page */
- continue;
- }
if (p->pg_flags & PG_BUSY) {
- simple_unlock(&uobj->vmobjlock);
uvmexp.pdbusy++;
/* someone else owns page, skip it */
continue;
@@ -514,9 +504,7 @@ uvmpd_scan_inactive(struct pglist *pglst)
if (p->pg_flags & PG_CLEAN) {
if (p->pg_flags & PQ_SWAPBACKED) {
/* this page now lives only in swap */
- simple_lock(&uvm.swap_data_lock);
uvmexp.swpgonly++;
- simple_unlock(&uvm.swap_data_lock);
}
/* zap all mappings with pmap_page_protect... */
@@ -535,11 +523,6 @@ uvmpd_scan_inactive(struct pglist *pglst)
/* remove from object */
anon->an_page = NULL;
- simple_unlock(&anon->an_lock);
- } else {
- /* pagefree has already removed the
- * page from the object */
- simple_unlock(&uobj->vmobjlock);
}
continue;
}
@@ -550,11 +533,6 @@ uvmpd_scan_inactive(struct pglist *pglst)
*/
if (free + uvmexp.paging > uvmexp.freetarg << 2) {
- if (anon) {
- simple_unlock(&anon->an_lock);
- } else {
- simple_unlock(&uobj->vmobjlock);
- }
continue;
}
@@ -570,11 +548,6 @@ uvmpd_scan_inactive(struct pglist *pglst)
uvmexp.swpgonly == uvmexp.swpages) {
dirtyreacts++;
uvm_pageactivate(p);
- if (anon) {
- simple_unlock(&anon->an_lock);
- } else {
- simple_unlock(&uobj->vmobjlock);
- }
continue;
}
@@ -650,12 +623,6 @@ uvmpd_scan_inactive(struct pglist *pglst)
&p->pg_flags,
PG_BUSY);
UVM_PAGE_OWN(p, NULL);
- if (anon)
- simple_unlock(
- &anon->an_lock);
- else
- simple_unlock(
- &uobj->vmobjlock);
continue;
}
swcpages = 0; /* cluster is empty */
@@ -692,11 +659,6 @@ uvmpd_scan_inactive(struct pglist *pglst)
if (swap_backed) {
if (p) { /* if we just added a page to cluster */
- if (anon)
- simple_unlock(&anon->an_lock);
- else
- simple_unlock(&uobj->vmobjlock);
-
/* cluster not full yet? */
if (swcpages < swnpages)
continue;
@@ -823,14 +785,6 @@ uvmpd_scan_inactive(struct pglist *pglst)
/* relock p's object: page queues not lock yet, so
* no need for "try" */
- /* !swap_backed case: already locked... */
- if (swap_backed) {
- if (anon)
- simple_lock(&anon->an_lock);
- else
- simple_lock(&uobj->vmobjlock);
- }
-
#ifdef DIAGNOSTIC
if (result == VM_PAGER_UNLOCK)
panic("pagedaemon: pageout returned "
@@ -856,7 +810,6 @@ uvmpd_scan_inactive(struct pglist *pglst)
anon->an_page = NULL;
p->uanon = NULL;
- simple_unlock(&anon->an_lock);
uvm_anfree(anon); /* kills anon */
pmap_page_protect(p, VM_PROT_NONE);
anon = NULL;
@@ -891,11 +844,6 @@ uvmpd_scan_inactive(struct pglist *pglst)
* true for active queue]).
*/
- if (anon)
- simple_unlock(&anon->an_lock);
- else if (uobj)
- simple_unlock(&uobj->vmobjlock);
-
if (nextpg && (nextpg->pg_flags & PQ_INACTIVE) == 0) {
nextpg = TAILQ_FIRST(pglst); /* reload! */
}
@@ -1009,8 +957,6 @@ uvmpd_scan(void)
/* is page anon owned or ownerless? */
if ((p->pg_flags & PQ_ANON) || p->uobject == NULL) {
KASSERT(p->uanon != NULL);
- if (!simple_lock_try(&p->uanon->an_lock))
- continue;
/* take over the page? */
if ((p->pg_flags & PQ_ANON) == 0) {
@@ -1018,9 +964,6 @@ uvmpd_scan(void)
p->loan_count--;
atomic_setbits_int(&p->pg_flags, PQ_ANON);
}
- } else {
- if (!simple_lock_try(&p->uobject->vmobjlock))
- continue;
}
/*
@@ -1028,10 +971,6 @@ uvmpd_scan(void)
*/
if ((p->pg_flags & PG_BUSY) != 0) {
- if (p->pg_flags & PQ_ANON)
- simple_unlock(&p->uanon->an_lock);
- else
- simple_unlock(&p->uobject->vmobjlock);
continue;
}
@@ -1071,9 +1010,5 @@ uvmpd_scan(void)
uvmexp.pddeact++;
inactive_shortage--;
}
- if (p->pg_flags & PQ_ANON)
- simple_unlock(&p->uanon->an_lock);
- else
- simple_unlock(&p->uobject->vmobjlock);
}
}
diff --git a/sys/uvm/uvm_swap.c b/sys/uvm/uvm_swap.c
index cf377792e29..54be0771b3d 100644
--- a/sys/uvm/uvm_swap.c
+++ b/sys/uvm/uvm_swap.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: uvm_swap.c,v 1.113 2013/05/03 13:57:46 florian Exp $ */
+/* $OpenBSD: uvm_swap.c,v 1.114 2013/05/30 15:17:59 tedu Exp $ */
/* $NetBSD: uvm_swap.c,v 1.40 2000/11/17 11:39:39 mrg Exp $ */
/*
@@ -92,9 +92,6 @@
* - swap_syscall_lock (sleep lock): this lock serializes the swapctl
* system call and prevents the swap priority list from changing
* while we are in the middle of a system call (e.g. SWAP_STATS).
- * - uvm.swap_data_lock (simple_lock): this lock protects all swap data
- * structures including the priority list, the swapdev structures,
- * and the swapmap extent.
*
* each swap device has the following info:
* - swap device in use (could be disabled, preventing future use)
@@ -285,7 +282,6 @@ uvm_swap_init(void)
LIST_INIT(&swap_priority);
uvmexp.nswapdev = 0;
- simple_lock_init(&uvm.swap_data_lock);
if (!swapdev_vp && bdevvp(swapdev, &swapdev_vp))
panic("uvm_swap_init: can't get vnode for swap device");
@@ -330,7 +326,6 @@ uvm_swap_initcrypt_all(void)
struct swappri *spp;
int npages;
- simple_lock(&uvm.swap_data_lock);
LIST_FOREACH(spp, &swap_priority, spi_swappri) {
CIRCLEQ_FOREACH(sdp, &spp->spi_swapdev, swd_next)
@@ -340,7 +335,6 @@ uvm_swap_initcrypt_all(void)
uvm_swap_initcrypt(sdp, npages);
}
}
- simple_unlock(&uvm.swap_data_lock);
}
void
@@ -455,8 +449,6 @@ uvm_swap_finicrypt_all(void)
struct swap_key *key;
unsigned int nkeys;
- simple_lock(&uvm.swap_data_lock);
-
LIST_FOREACH(spp, &swap_priority, spi_swappri) {
CIRCLEQ_FOREACH(sdp, &spp->spi_swapdev, swd_next) {
if (sdp->swd_decrypt == NULL)
@@ -470,7 +462,6 @@ uvm_swap_finicrypt_all(void)
} while (key-- != sdp->swd_keys);
}
}
- simple_unlock(&uvm.swap_data_lock);
}
#endif /* UVM_SWAP_ENCRYPT */
@@ -751,14 +742,12 @@ sys_swapctl(struct proc *p, void *v, register_t *retval)
*/
priority = SCARG(uap, misc);
spp = malloc(sizeof *spp, M_VMSWAP, M_WAITOK);
- simple_lock(&uvm.swap_data_lock);
if ((sdp = swaplist_find(vp, 1)) == NULL) {
error = ENOENT;
} else {
swaplist_insert(sdp, spp, priority);
swaplist_trim();
}
- simple_unlock(&uvm.swap_data_lock);
if (error)
free(spp, M_VMSWAP);
break;
@@ -773,10 +762,8 @@ sys_swapctl(struct proc *p, void *v, register_t *retval)
*/
priority = SCARG(uap, misc);
- simple_lock(&uvm.swap_data_lock);
if ((sdp = swaplist_find(vp, 0)) != NULL) {
error = EBUSY;
- simple_unlock(&uvm.swap_data_lock);
break;
}
sdp = malloc(sizeof *sdp, M_VMSWAP, M_WAITOK|M_ZERO);
@@ -793,7 +780,6 @@ sys_swapctl(struct proc *p, void *v, register_t *retval)
}
swaplist_insert(sdp, spp, priority);
- simple_unlock(&uvm.swap_data_lock);
sdp->swd_pathlen = len;
sdp->swd_path = malloc(sdp->swd_pathlen, M_VMSWAP, M_WAITOK);
@@ -808,10 +794,8 @@ sys_swapctl(struct proc *p, void *v, register_t *retval)
*/
if ((error = swap_on(p, sdp)) != 0) {
- simple_lock(&uvm.swap_data_lock);
(void) swaplist_find(vp, 1); /* kill fake entry */
swaplist_trim();
- simple_unlock(&uvm.swap_data_lock);
if (vp->v_type == VREG) {
crfree(sdp->swd_cred);
}
@@ -822,9 +806,7 @@ sys_swapctl(struct proc *p, void *v, register_t *retval)
break;
case SWAP_OFF:
- simple_lock(&uvm.swap_data_lock);
if ((sdp = swaplist_find(vp, 0)) == NULL) {
- simple_unlock(&uvm.swap_data_lock);
error = ENXIO;
break;
}
@@ -834,7 +816,6 @@ sys_swapctl(struct proc *p, void *v, register_t *retval)
* can't stop swapping from it (again).
*/
if ((sdp->swd_flags & (SWF_INUSE|SWF_ENABLE)) == 0) {
- simple_unlock(&uvm.swap_data_lock);
error = EBUSY;
break;
}
@@ -1016,13 +997,11 @@ swap_on(struct proc *p, struct swapdev *sdp)
/*
* now add the new swapdev to the drum and enable.
*/
- simple_lock(&uvm.swap_data_lock);
swapdrum_add(sdp, npages);
sdp->swd_npages = size;
sdp->swd_flags &= ~SWF_FAKE; /* going live */
sdp->swd_flags |= (SWF_INUSE|SWF_ENABLE);
uvmexp.swpages += size;
- simple_unlock(&uvm.swap_data_lock);
return (0);
bad:
@@ -1046,7 +1025,6 @@ swap_off(struct proc *p, struct swapdev *sdp)
/* disable the swap area being removed */
sdp->swd_flags &= ~SWF_ENABLE;
- simple_unlock(&uvm.swap_data_lock);
/*
* the idea is to find all the pages that are paged out to this
@@ -1066,9 +1044,7 @@ swap_off(struct proc *p, struct swapdev *sdp)
}
if (error) {
- simple_lock(&uvm.swap_data_lock);
sdp->swd_flags |= SWF_ENABLE;
- simple_unlock(&uvm.swap_data_lock);
return (error);
}
@@ -1085,7 +1061,6 @@ swap_off(struct proc *p, struct swapdev *sdp)
(void) VOP_CLOSE(sdp->swd_vp, FREAD|FWRITE, p->p_ucred, p);
}
- simple_lock(&uvm.swap_data_lock);
uvmexp.swpages -= sdp->swd_npages;
if (swaplist_find(sdp->swd_vp, 1) == NULL)
@@ -1099,7 +1074,6 @@ swap_off(struct proc *p, struct swapdev *sdp)
EX_WAITOK);
extent_destroy(sdp->swd_ex);
free(sdp, M_VMSWAP);
- simple_unlock(&uvm.swap_data_lock);
return (0);
}
@@ -1124,9 +1098,7 @@ swstrategy(struct buf *bp)
* in it (i.e. the blocks we are doing I/O on).
*/
pageno = dbtob((u_int64_t)bp->b_blkno) >> PAGE_SHIFT;
- simple_lock(&uvm.swap_data_lock);
sdp = swapdrum_getsdp(pageno);
- simple_unlock(&uvm.swap_data_lock);
if (sdp == NULL) {
bp->b_error = EINVAL;
bp->b_flags |= B_ERROR;
@@ -1481,7 +1453,6 @@ uvm_swap_alloc(int *nslots, boolean_t lessok)
/*
* lock data lock, convert slots into blocks, and enter loop
*/
- simple_lock(&uvm.swap_data_lock);
ReTry: /* XXXMRG */
for (spp = LIST_FIRST(&swap_priority); spp != NULL;
@@ -1507,7 +1478,6 @@ ReTry: /* XXXMRG */
CIRCLEQ_INSERT_TAIL(&spp->spi_swapdev, sdp, swd_next);
sdp->swd_npginuse += *nslots;
uvmexp.swpginuse += *nslots;
- simple_unlock(&uvm.swap_data_lock);
/* done! return drum slot number */
return(result + sdp->swd_drumoffset);
}
@@ -1520,7 +1490,6 @@ ReTry: /* XXXMRG */
}
/* XXXMRG: END HACK */
- simple_unlock(&uvm.swap_data_lock);
return 0; /* failed */
}
@@ -1534,7 +1503,6 @@ uvm_swap_markbad(int startslot, int nslots)
{
struct swapdev *sdp;
- simple_lock(&uvm.swap_data_lock);
sdp = swapdrum_getsdp(startslot);
if (sdp != NULL) {
/*
@@ -1545,7 +1513,6 @@ uvm_swap_markbad(int startslot, int nslots)
*/
sdp->swd_npgbad += nslots;
}
- simple_unlock(&uvm.swap_data_lock);
}
/*
@@ -1573,7 +1540,6 @@ uvm_swap_free(int startslot, int nslots)
* lookup and access the extent.
*/
- simple_lock(&uvm.swap_data_lock);
sdp = swapdrum_getsdp(startslot);
KASSERT(uvmexp.nswapdev >= 1);
KASSERT(sdp != NULL);
@@ -1605,7 +1571,6 @@ uvm_swap_free(int startslot, int nslots)
}
}
#endif /* UVM_SWAP_ENCRYPT */
- simple_unlock(&uvm.swap_data_lock);
}
/*
@@ -1645,9 +1610,7 @@ uvm_swap_get(struct vm_page *page, int swslot, int flags)
/*
* this page is (about to be) no longer only in swap.
*/
- simple_lock(&uvm.swap_data_lock);
uvmexp.swpgonly--;
- simple_unlock(&uvm.swap_data_lock);
result = uvm_swap_io(&page, swslot, 1, B_READ |
((flags & PGO_SYNCIO) ? 0 : B_ASYNC));
@@ -1656,9 +1619,7 @@ uvm_swap_get(struct vm_page *page, int swslot, int flags)
/*
* oops, the read failed so it really is still only in swap.
*/
- simple_lock(&uvm.swap_data_lock);
uvmexp.swpgonly++;
- simple_unlock(&uvm.swap_data_lock);
}
return (result);
@@ -1724,9 +1685,7 @@ uvm_swap_io(struct vm_page **pps, int startslot, int npages, int flags)
* XXX - does this information stay the same over the whole
* execution of this function?
*/
- simple_lock(&uvm.swap_data_lock);
sdp = swapdrum_getsdp(startslot);
- simple_unlock(&uvm.swap_data_lock);
}
/*
diff --git a/sys/uvm/uvm_vnode.c b/sys/uvm/uvm_vnode.c
index 130cf88ae0c..36691f59906 100644
--- a/sys/uvm/uvm_vnode.c
+++ b/sys/uvm/uvm_vnode.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: uvm_vnode.c,v 1.76 2011/07/04 20:35:35 deraadt Exp $ */
+/* $OpenBSD: uvm_vnode.c,v 1.77 2013/05/30 15:17:59 tedu Exp $ */
/* $NetBSD: uvm_vnode.c,v 1.36 2000/11/24 20:34:01 chs Exp $ */
/*
@@ -159,19 +159,16 @@ uvn_attach(void *arg, vm_prot_t accessprot)
/*
* first get a lock on the uvn.
*/
- simple_lock(&uvn->u_obj.vmobjlock);
while (uvn->u_flags & UVM_VNODE_BLOCKED) {
uvn->u_flags |= UVM_VNODE_WANTED;
UVM_UNLOCK_AND_WAIT(uvn, &uvn->u_obj.vmobjlock, FALSE,
"uvn_attach", 0);
- simple_lock(&uvn->u_obj.vmobjlock);
}
/*
* if we're mapping a BLK device, make sure it is a disk.
*/
if (vp->v_type == VBLK && bdevsw[major(vp->v_rdev)].d_type != D_DISK) {
- simple_unlock(&uvn->u_obj.vmobjlock); /* drop lock */
return(NULL);
}
@@ -198,7 +195,6 @@ uvn_attach(void *arg, vm_prot_t accessprot)
}
/* unlock and return */
- simple_unlock(&uvn->u_obj.vmobjlock);
return (&uvn->u_obj);
}
@@ -211,8 +207,6 @@ uvn_attach(void *arg, vm_prot_t accessprot)
* it.
*/
uvn->u_flags = UVM_VNODE_ALOCK;
- simple_unlock(&uvn->u_obj.vmobjlock); /* drop lock in case we sleep */
- /* XXX: curproc? */
if (vp->v_type == VBLK) {
/*
@@ -237,13 +231,11 @@ uvn_attach(void *arg, vm_prot_t accessprot)
}
/* relock object */
- simple_lock(&uvn->u_obj.vmobjlock);
if (result != 0) {
if (uvn->u_flags & UVM_VNODE_WANTED)
wakeup(uvn);
uvn->u_flags = 0;
- simple_unlock(&uvn->u_obj.vmobjlock); /* drop lock */
return(NULL);
}
@@ -277,7 +269,6 @@ uvn_attach(void *arg, vm_prot_t accessprot)
* reference count goes to zero [and we either free or persist].
*/
vref(vp);
- simple_unlock(&uvn->u_obj.vmobjlock);
if (oldflags & UVM_VNODE_WANTED)
wakeup(uvn);
@@ -304,7 +295,6 @@ uvn_reference(struct uvm_object *uobj)
struct uvm_vnode *uvn = (struct uvm_vnode *) uobj;
#endif
- simple_lock(&uobj->vmobjlock);
#ifdef DEBUG
if ((uvn->u_flags & UVM_VNODE_VALID) == 0) {
printf("uvn_reference: ref=%d, flags=0x%x\n", uvn->u_flags,
@@ -313,7 +303,6 @@ uvn_reference(struct uvm_object *uobj)
}
#endif
uobj->uo_refs++;
- simple_unlock(&uobj->vmobjlock);
}
/*
@@ -332,11 +321,9 @@ uvn_detach(struct uvm_object *uobj)
struct vnode *vp;
int oldflags;
- simple_lock(&uobj->vmobjlock);
uobj->uo_refs--; /* drop ref! */
if (uobj->uo_refs) { /* still more refs */
- simple_unlock(&uobj->vmobjlock);
return;
}
@@ -361,7 +348,6 @@ uvn_detach(struct uvm_object *uobj)
if (uvn->u_flags & UVM_VNODE_CANPERSIST) {
/* won't block */
uvn_flush(uobj, 0, 0, PGO_DEACTIVATE|PGO_ALLPAGES);
- simple_unlock(&uobj->vmobjlock);
vrele(vp); /* drop vnode reference */
return;
}
@@ -399,7 +385,6 @@ uvn_detach(struct uvm_object *uobj)
uvn->u_flags |= UVM_VNODE_IOSYNC;
UVM_UNLOCK_AND_WAIT(&uvn->u_nio, &uvn->u_obj.vmobjlock, FALSE,
"uvn_term",0);
- simple_lock(&uvn->u_obj.vmobjlock);
}
if ((uvn->u_flags & UVM_VNODE_RELKILL) == 0)
@@ -415,7 +400,6 @@ uvn_detach(struct uvm_object *uobj)
KASSERT(RB_EMPTY(&uobj->memt));
oldflags = uvn->u_flags;
uvn->u_flags = 0;
- simple_unlock(&uobj->vmobjlock);
/* wake up any sleepers */
if (oldflags & UVM_VNODE_WANTED)
@@ -464,9 +448,7 @@ uvm_vnp_terminate(struct vnode *vp)
/*
* lock object and check if it is valid
*/
- simple_lock(&uvn->u_obj.vmobjlock);
if ((uvn->u_flags & UVM_VNODE_VALID) == 0) {
- simple_unlock(&uvn->u_obj.vmobjlock);
return;
}
@@ -539,7 +521,6 @@ uvm_vnp_terminate(struct vnode *vp)
uvn->u_flags |= UVM_VNODE_IOSYNC;
UVM_UNLOCK_AND_WAIT(&uvn->u_nio, &uvn->u_obj.vmobjlock, FALSE,
"uvn_term",0);
- simple_lock(&uvn->u_obj.vmobjlock);
}
/*
@@ -578,7 +559,6 @@ uvm_vnp_terminate(struct vnode *vp)
if (oldflags & UVM_VNODE_WANTED)
wakeup(uvn); /* object lock still held */
- simple_unlock(&uvn->u_obj.vmobjlock);
}
@@ -756,7 +736,6 @@ uvn_flush(struct uvm_object *uobj, voff_t start, voff_t stop, int flags)
uvm_unlock_pageq();
UVM_UNLOCK_AND_WAIT(pp,
&uobj->vmobjlock, 0, "uvn_flsh", 0);
- simple_lock(&uobj->vmobjlock);
uvm_lock_pageq();
curoff -= PAGE_SIZE;
continue;
@@ -802,7 +781,6 @@ ReTry:
*/
/* relock! */
- simple_lock(&uobj->vmobjlock);
uvm_lock_pageq();
/*
@@ -926,7 +904,6 @@ ReTry:
uvn->u_flags |= UVM_VNODE_IOSYNC;
UVM_UNLOCK_AND_WAIT(&uvn->u_nio, &uvn->u_obj.vmobjlock,
FALSE, "uvn_flush",0);
- simple_lock(&uvn->u_obj.vmobjlock);
}
if (uvn->u_flags & UVM_VNODE_IOSYNCWANTED)
wakeup(&uvn->u_flags);
@@ -1134,9 +1111,7 @@ uvn_get(struct uvm_object *uobj, voff_t offset, struct vm_page **pps,
/* out of RAM? */
if (ptmp == NULL) {
- simple_unlock(&uobj->vmobjlock);
uvm_wait("uvn_getpage");
- simple_lock(&uobj->vmobjlock);
/* goto top of pps while loop */
continue;
@@ -1154,7 +1129,6 @@ uvn_get(struct uvm_object *uobj, voff_t offset, struct vm_page **pps,
atomic_setbits_int(&ptmp->pg_flags, PG_WANTED);
UVM_UNLOCK_AND_WAIT(ptmp,
&uobj->vmobjlock, FALSE, "uvn_get",0);
- simple_lock(&uobj->vmobjlock);
continue; /* goto top of pps while loop */
}
@@ -1194,7 +1168,6 @@ uvn_get(struct uvm_object *uobj, voff_t offset, struct vm_page **pps,
*/
/* lock object. check for errors. */
- simple_lock(&uobj->vmobjlock);
if (result != VM_PAGER_OK) {
if (ptmp->pg_flags & PG_WANTED)
/* object lock still held */
@@ -1206,7 +1179,6 @@ uvn_get(struct uvm_object *uobj, voff_t offset, struct vm_page **pps,
uvm_lock_pageq();
uvm_pagefree(ptmp);
uvm_unlock_pageq();
- simple_unlock(&uobj->vmobjlock);
return(result);
}
@@ -1232,7 +1204,6 @@ uvn_get(struct uvm_object *uobj, voff_t offset, struct vm_page **pps,
* finally, unlock object and return.
*/
- simple_unlock(&uobj->vmobjlock);
return (VM_PAGER_OK);
}
@@ -1271,13 +1242,11 @@ uvn_io(struct uvm_vnode *uvn, vm_page_t *pps, int npages, int flags, int rw)
while (uvn->u_flags & UVM_VNODE_IOSYNC) {
if (waitf == M_NOWAIT) {
- simple_unlock(&uvn->u_obj.vmobjlock);
return(VM_PAGER_AGAIN);
}
uvn->u_flags |= UVM_VNODE_IOSYNCWANTED;
UVM_UNLOCK_AND_WAIT(&uvn->u_flags, &uvn->u_obj.vmobjlock,
FALSE, "uvn_iosync",0);
- simple_lock(&uvn->u_obj.vmobjlock);
}
/*
@@ -1285,7 +1254,6 @@ uvn_io(struct uvm_vnode *uvn, vm_page_t *pps, int npages, int flags, int rw)
*/
if (file_offset >= uvn->u_size) {
- simple_unlock(&uvn->u_obj.vmobjlock);
return(VM_PAGER_BAD);
}
@@ -1298,7 +1266,6 @@ uvn_io(struct uvm_vnode *uvn, vm_page_t *pps, int npages, int flags, int rw)
kva = uvm_pagermapin(pps, npages, mapinflags);
if (kva == 0 && waitf == M_NOWAIT) {
- simple_unlock(&uvn->u_obj.vmobjlock);
return(VM_PAGER_AGAIN);
}
@@ -1309,7 +1276,6 @@ uvn_io(struct uvm_vnode *uvn, vm_page_t *pps, int npages, int flags, int rw)
*/
uvn->u_nio++; /* we have an I/O in progress! */
- simple_unlock(&uvn->u_obj.vmobjlock);
/* NOTE: object now unlocked */
if (kva == 0)
kva = uvm_pagermapin(pps, npages,
@@ -1395,14 +1361,12 @@ uvn_io(struct uvm_vnode *uvn, vm_page_t *pps, int npages, int flags, int rw)
* now clean up the object (i.e. drop I/O count)
*/
- simple_lock(&uvn->u_obj.vmobjlock);
/* NOTE: object now locked! */
uvn->u_nio--; /* I/O DONE! */
if ((uvn->u_flags & UVM_VNODE_IOSYNC) != 0 && uvn->u_nio == 0) {
wakeup(&uvn->u_nio);
}
- simple_unlock(&uvn->u_obj.vmobjlock);
/* NOTE: object now unlocked! */
/*
@@ -1460,10 +1424,8 @@ uvm_vnp_uncache(struct vnode *vp)
* lock uvn part of the vnode and check to see if we need to do anything
*/
- simple_lock(&uvn->u_obj.vmobjlock);
if ((uvn->u_flags & UVM_VNODE_VALID) == 0 ||
(uvn->u_flags & UVM_VNODE_BLOCKED) != 0) {
- simple_unlock(&uvn->u_obj.vmobjlock);
return(TRUE);
}
@@ -1474,7 +1436,6 @@ uvm_vnp_uncache(struct vnode *vp)
uvn->u_flags &= ~UVM_VNODE_CANPERSIST;
if (uvn->u_obj.uo_refs) {
- simple_unlock(&uvn->u_obj.vmobjlock);
return(FALSE);
}
@@ -1485,7 +1446,6 @@ uvm_vnp_uncache(struct vnode *vp)
vref(vp); /* seems ok, even with VOP_LOCK */
uvn->u_obj.uo_refs++; /* value is now 1 */
- simple_unlock(&uvn->u_obj.vmobjlock);
#ifdef VFSLCKDEBUG
/*
@@ -1543,7 +1503,6 @@ uvm_vnp_setsize(struct vnode *vp, voff_t newsize)
/*
* lock uvn and check for valid object, and if valid: do it!
*/
- simple_lock(&uvn->u_obj.vmobjlock);
if (uvn->u_flags & UVM_VNODE_VALID) {
/*
@@ -1557,7 +1516,6 @@ uvm_vnp_setsize(struct vnode *vp, voff_t newsize)
}
uvn->u_size = newsize;
}
- simple_unlock(&uvn->u_obj.vmobjlock);
/*
* done