Age | Commit message (Collapse) | Author |
|
|
|
the big change is removing the integration with and reliance on
bridge(4) for learning vxlan endpoints. we have the etherbridge
layer now (which is used by veb, nvgre, bpe, etc) so vxlan can
operate independently of bridge(4) (or any other driver) while still
dynamically learning about other endpoints.
vxlan now uses the udp socket upcall mechanism to receive packets.
this means it actually creates and binds udp sockets to use rather
adding code in the udp layer for stealing packets from the udp
layer.
i think it's also important to note that this adds loop prevention
to the code. this stops a vxlan interface being used to transmit a
packet that was encapsulated in itself.
i want to clear this out of my tree where it's been sitting for
nearly a year. noone seems too concerned with the change either
way.
ok claudio@
|
|
Ok deraat@
|
|
ok kettenis@
|
|
with AML parsing outside the acpi thread, the locking-release dance
around wsdisplay_{suspend,resume} can be removed
ok kettenis
|
|
ok deraadt@
|
|
reset the MD state before bailing out. New MD function sleep_abort()
does that.
|
|
Like set-clipboard and allow-rename it is safer to forbid this by
default.
|
|
query.
|
|
Fixes my com0 issues after zzz and ZZZ.
OK kettenis@ miod@
|
|
|
|
The (known) lock order reversals which now occur more reliably and much
earlier on WITNESS boots with this diff knock out syzcaller reports since
syzcaller stops at the first "crash report":
https://syzkaller.appspot.com/bug?id=81b39e970cd2eb21b97d1b31746c693e300fd2dd
|
|
ok miod@
|
|
|
|
OK dlg@
|
|
|
|
okay afresh1@
|
|
have a state once I'm past the initial locking
|
|
spotted by HARUYAMA Seigo
|
|
tun_dev_open and tun_dev_close were being optmistic.
|
|
it only works on struct ifnet data, so passing ifp makes it clearer
what's actually being manipulated. also fix tun_dev_open so
tun_link_state is called before if_put instead of immediately after.
|
|
ok deraadt@ guenther@
|
|
posix requires a byte to be 8 bits
|
|
in sleep_resume(), which seems sensible for other future systems also
|
|
steps, so shuffle the MD steps to reinforce the order.
|
|
inst_trap_return() was checking for rfir by masking with 0xfc001fc0
which made it impossible to match rfir (0xca0)
rfi: return from interruption
00 rv rv rv 60 0
6 5 5 3 8 5
rfir: return from interruption and restore
00 rv rv rv 65 0
6 5 5 3 8 5
from "PA-RISC 1.1 Architecture and Instruction Set Reference Manual"
where rv indicates reserved bits
change the mask to only mask out reserved bits and check for rfi
in addition to rfir
ok miod@
|
|
accidentally broken with the reorganization of the suspend/resume code that
was committed a few days ago.
ok deraadt@
|
|
ok miod@
|
|
ok claudio@ deraadt@
|
|
rather than relying on /usr/mdec/mbr.
Tested by tobhe@
|
|
This is an updated version of uvm_map.c r1.283 "Unwire with map lock held".
The previous version introduced a use-after-free by not unlocking vm_map
locks in uvm_map_teardown(), resulting in dangling references on the
reaper's lock list (thanks visa!).
Lock and unlock the map in around uvm_map_teardown() instead.
This code path holds the last reference, hence the lock isn't strictly
needed except for satisfying upcoming locking assertions.
Tested on amd64, arm64, i386, macppc, octeon, sparc64.
This time also with WITNESS enabled (except on sparc64 which builds but does
not boot with WITNESS; this is a known issue).
OK mpi visa
|
|
|
|
Ok deraadt
|
|
|
|
match the device.
ok visa@
|
|
ok visa@
|
|
OK claudio@ tb@
|
|
OK claudio@ tb@
|
|
This prevents errors on luna88k in src/usr.sbin/fw_update/pattern.c
which includes dev/pci/pcivar.h, which includes machine/bus.h.
Suggested by miod@
|
|
spotted by and ok tb@, ok tobhe@
|
|
from cursig() to postsig() or the caller itself. This will simplify locking.
Also alter sigactsfree() a bit and move it into process_zap() so ps_sigacts
is always a valid pointer.
OK semarie@
|
|
|
|
|
|
ok miod@
|
|
|
|
|
|
adapted from a diff by Ted Bullock
|
|
|
|
from Ted Bullock who tested on sparc64 with xvr-100
|
|
previously sbchecklowmem() (and sonewconn()) would look at the mbuf
and mbuf cluster pools to see if they were approaching their hard
limits. based on how many mbufs/clusters were allocated against the
limits, socket operations would start to fail with ENOBUFS until
utilisation went down.
mbufs and clusters have changed a lot since then though. there are
now many mbuf cluster pools, not just one for 2k clusters. because
of this the mbuf layer now limits the amount of memory all the mbuf
pools can allocate backend pages from rather than limit the individual
pools. this means sbchecklowmem() ends up looking at the default
pool hard limit, which is UINT_MAX, which in turn means means
sbchecklowmem() probably never applies backpressure. this is made
worse on multiprocessor systems where per cpu caches of mbuf and
cluster pool items are enabled because the number of in use pool
items is distorted by the cpu caches.
this switches sbchecklowmem to looking at the page allocations made
by all the pools instead. the big benefit of this is that the page
allocations are much more representative of the overall mbuf memory
usage in the system. the downside is is that the backend page
allocation accounting does not see idle memory held by pools. pools
cannot release partially free pages to the page backend (obviously),
and pools cache idle items to avoid thrashing on the backend page
allocator. this means the page allocation level is higher than the
memory used by actual in-flight mbufs.
however, this can also be a benefit. the backend page allocation is a
kind of smoothed out "trend" line. mbuf utilisation over short periods
can be extremely bursty because of things like rx ring dequeue and fill
cycles, or large socket sends. if you're trying to grow socket
buffers while these things are happening, luck becomes an important
factor in whether it will work or not. because pools cache idle items,
the backend page utilisation better represents the overall trend
of activity in the system and will give more consistent behaviour here.
this diff is deliberately simple. we're basically going from "no
limits" to "some sort of limit" for sockets again, so keeping the
code simple means it should be easy to understand and tweak in the
future.
ok djm@ visa@ claudio@
|