Age | Commit message (Collapse) | Author |
|
Okay weingart@, "I'm game with putting my name on it" dlg@
|
|
no change for normal code
|
|
miod@ ok
|
|
us did not see it or get a chance to test it before it was commited. It
broke cvs, in the ami driver, making it not succeed at seeing it's devices.
|
|
proper; found by krause and mmap_fixed
|
|
this results in lesse kva waste due to static preallocation of those
for every phys page and also every swap page.
tested by beck krw miod
|
|
|
|
|
|
will prevent panics in, e.g., bus_dmamem_alloc().
ok jason@ art@
|
|
defined; from NetBSD. Currently only used on xscale arm to use the mini data
cache for u area mappings instead of the main data cache.
|
|
|
|
1. drain hooks and lists of allocators make the code complicated
2. the only hooks in the system are the mbuf reclaim routines
3. if reclaim is actually able to put a meaningful amount of memory back
in the system, i think something else is dicked up. ie, if reclaiming
your ip fragment buffers makes the difference thrashing swap and not,
your system is in a load of trouble.
4. it's a scary amount of code running with very weird spl requirements
and i'd say it's pretty much totally untested. raise your hand if your
router is running at the edge of swap.
5. the reclaim stuff goes back to when mbufs lived in a tiny vm_map and
you could run out of va. that's very unlikely (like impossible) now.
ok/tested pedro krw sturm
|
|
as paddr_t could be a long long (soon) always cast and print as llx.
|
|
as freepages being vconverted back to byte address make sure to
perform calculations in (upcoming) larger paddr_t to avoid losing
higher bits in calculation.
|
|
physmem); miod@ toby@ ok
|
|
call uvm_unmap_p instead of uvm_unmap so that it has the process information
and can adjust vm_dused. okay pedro@ tedu@
|
|
page, so the arithmetic was ok. Spotted by david@
|
|
uvm_vsunlock(). ok mickey@
|
|
|
|
kettenis@ tedu@ ok
|
|
pgfree check into pglist; no functional change for normal kernels; make histories uncommon
|
|
ok deraadt@
|
|
by lint.
"Probably a bogus cut'n paste." says moid.
ok miod@ pedro@
|
|
|
|
14060 skip MADV_SEQUENTIAL if refaulting
18037 missing pageactivate
tested for some time by jolan krw
|
|
miod@ mickey@
|
|
of panics and bugfixes. Access curproc directly, do not expect a process
pointer as an argument. Should fix many "process context required" bugs.
Incentive and okay millert@, okay marc@. Various testing, thanks.
|
|
|
|
the protection of the memory mapping we're doing I/O on, or if we want to
leave them as they are. This should only be necessary for breakpoint
insertion in code, so we'll only use it for ptrace requests.
Initially from art@ after discussion with kettenis@ millert@ and I,
tested by many.
|
|
done in uvm_swapin(). Looks like this was a mistake made while editing. No
response from art@. deraadt@, miod@, pedro@ ok
|
|
practise. Depending on the list implementation, this might or might
not work; so make it use a safe idiom. ok pedro@ millert@ deraadt@
|
|
|
|
|
|
a proper fix has been implemented in uvm_mapent_alloc().
ok pedro@
|
|
|
|
the uvm_km_page allocator and use it instead of calling panic()
- add a counter to uvmexp so we can keep track of how many map entries
we have in use
idea from tedu@, long ago, okay deraadt@
|
|
arches; except on sparc where the range is 4-8 for !sun4m and 4-64 for sun4m,
selected at runtime.
|
|
|
|
- Use it to skip device mappings while dumping core.
- Ignore EFAULT errors while dumping core since they can happen
even for valid mappings. Just skip that part of the core file and
let it get automagically zero-filled.
This fixes the broken X core dumps that people have been seeing and also
fixes some other potential problems that could prevent core dumps (mmaps
beyond EOF, etc.).
tedu@ ok
|
|
|
|
over past week. as a bonus, kills 5 XXXs.
|
|
one case fixed here).
miod@ "appears to be harmless"
markus@ ok
|
|
|
|
everyone for the prompt review and ok of this work ;-) Yeah, that includes me
too, or maybe especially me. I am sorry.
Change the sched_lock to a mutex. This fixes, among other things, the infamous
"telnet localhost &" problem. The real bug in that case was that the sched_lock
which is by design a non-recursive lock, was recursively acquired, and not
enough releases made us hold the lock in the idle loop, blocking scheduling
on the other processors. Some of the other processors would hold the biglock though,
which made it impossible for cpu 0 to enter the kernel... A nice deadlock.
Let me just say debugging this for days just to realize that it was all fixed
in an old diff noone ever ok'd was somewhat of an anti-climax.
This diff also changes splsched to be correct for all our architectures.
|
|
pages a process uses. this is now the userland "data size" value.
ok art deraadt tdeval. thanks testers.
|
|
|
|
|
|
|
|
who have machines that hit swap a lot. decided after survey of developers,
we found that most turned this on. ok various
|
|
other archs for now, beck theo ok
|