Update FreeBSD support.

master
Fedor 2021-02-07 17:32:56 +02:00
parent 5ea204f926
commit 0b5dd854b9
6 changed files with 375 additions and 82 deletions

View File

@ -63,6 +63,11 @@ def GeckoBinary(linkage='dependent', msvcrt='dynamic', mozglue=None):
LDFLAGS += ['-rdynamic']
if CONFIG['MOZ_MEMORY']:
USE_LIBS += ['memory']
if CONFIG['OS_ARCH'] == 'FreeBSD':
# Make sure this function is linked in, so that it is
# executed at executable load (it has the 'constructor'
# flag).
LDFLAGS += ['-u', 'jemalloc_FreeBSD_init']
elif mozglue == 'library':
LIBRARY_DEFINES['MOZ_HAS_MOZGLUE'] = True
if not CONFIG['MOZ_GLUE_IN_PROGRAM']:

View File

@ -138,12 +138,3 @@ evutil_secure_rng_get_bytes(void *buf, size_t n)
{
ev_arc4random_buf(buf, n);
}
#if !defined(__OpenBSD__) && !defined(ANDROID) && !defined(__sun__)
void
evutil_secure_rng_add_bytes(const char *buf, size_t n)
{
arc4random_addrandom((unsigned char*)buf,
n>(size_t)INT_MAX ? INT_MAX : (int)n);
}
#endif

View File

@ -672,24 +672,6 @@ void evutil_secure_rng_get_bytes(void *buf, size_t n);
*/
int evutil_secure_rng_init(void);
#if !defined(__OpenBSD__) && !defined(ANDROID) && !defined(__sun__)
/** Seed the random number generator with extra random bytes.
You should almost never need to call this function; it should be
sufficient to invoke evutil_secure_rng_init(), or let Libevent take
care of calling evutil_secure_rng_init() on its own.
If you call this function as a _replacement_ for the regular
entropy sources, then you need to be sure that your input
contains a fairly large amount of strong entropy. Doing so is
notoriously hard: most people who try get it wrong. Watch out!
@param dat a buffer full of a strong source of random numbers
@param datlen the number of bytes to read from datlen
*/
void evutil_secure_rng_add_bytes(const char *dat, size_t datlen);
#endif
#ifdef __cplusplus
}
#endif

View File

@ -179,6 +179,9 @@
#endif
#include <sys/types.h>
#ifdef MOZ_MEMORY_BSD
#include <sys/sysctl.h>
#endif
#include <errno.h>
#include <stdlib.h>
@ -494,10 +497,10 @@ typedef pthread_mutex_t malloc_spinlock_t;
#endif
/* Set to true once the allocator has been initialized. */
static bool malloc_initialized = false;
static volatile bool malloc_initialized = false;
#if defined(MOZ_MEMORY_WINDOWS)
/* No init lock for Windows. */
#if defined(MOZ_MEMORY_WINDOWS) || defined(__FreeBSD__)
/* No init lock for Windows nor FreeBSD. */
#elif defined(MOZ_MEMORY_DARWIN)
static malloc_mutex_t init_lock = {OS_SPINLOCK_INIT};
#elif defined(MOZ_MEMORY_LINUX) && !defined(MOZ_MEMORY_ANDROID)
@ -1399,6 +1402,11 @@ int pthread_atfork(void (*)(void), void (*)(void), void(*)(void));
* cases.
*/
#ifdef __FreeBSD__
// If true, memory calls must be diverted to the bootstrap allocator
static __thread bool in_mutex_init = false;
#endif
static bool
malloc_mutex_init(malloc_mutex_t *mutex)
{
@ -1417,6 +1425,19 @@ malloc_mutex_init(malloc_mutex_t *mutex)
return (true);
}
pthread_mutexattr_destroy(&attr);
#elif defined(__FreeBSD__)
in_mutex_init = true;
*mutex = PTHREAD_ADAPTIVE_MUTEX_INITIALIZER_NP;
// Make sure necessary mutex memory is allocated right now, with
// 'in_mutex_init' set to true (allocations to be diverted to the
// bootstrap allocator). Also force multi-thread initialization in
// libthr (checked and performed in 'pthread_mutex_lock').
pthread_mutex_lock(mutex);
pthread_mutex_unlock(mutex);
in_mutex_init = false;
#else
if (pthread_mutex_init(mutex, NULL) != 0)
return (true);
@ -1471,6 +1492,8 @@ malloc_spin_init(malloc_spinlock_t *lock)
return (true);
}
pthread_mutexattr_destroy(&attr);
#elif defined(__FreeBSD__)
malloc_lock_init(lock);
#else
if (pthread_mutex_init(lock, NULL) != 0)
return (true);
@ -4755,25 +4778,7 @@ huge_dalloc(void *ptr)
* Platform-specific methods to determine the number of CPUs in a system.
* This will be used to determine the desired number of arenas.
*/
#ifdef MOZ_MEMORY_BSD
static inline unsigned
malloc_ncpus(void)
{
unsigned ret;
int mib[2];
size_t len;
mib[0] = CTL_HW;
mib[1] = HW_NCPU;
len = sizeof(ret);
if (sysctl(mib, 2, &ret, &len, (void *) 0, 0) == -1) {
/* Error. */
return (1);
}
return (ret);
}
#elif (defined(MOZ_MEMORY_LINUX))
#if (defined(MOZ_MEMORY_LINUX))
#include <fcntl.h>
static inline unsigned
@ -4845,8 +4850,7 @@ malloc_ncpus(void)
else
return (n);
}
#elif (defined(MOZ_MEMORY_SOLARIS))
#elif (defined(MOZ_MEMORY_SOLARIS) || defined(MOZ_MEMORY_BSD))
static inline unsigned
malloc_ncpus(void)
{
@ -4991,18 +4995,13 @@ malloc_print_stats(void)
}
}
/*
* FreeBSD's pthreads implementation calls malloc(3), so the malloc
* implementation has to take pains to avoid infinite recursion during
* initialization.
*/
#if (defined(MOZ_MEMORY_WINDOWS) || defined(MOZ_MEMORY_DARWIN))
#define malloc_init() false
#else
static inline bool
malloc_init(void)
{
if (malloc_initialized == false)
return (malloc_init_hard());
@ -5010,6 +5009,238 @@ malloc_init(void)
}
#endif
#ifdef __FreeBSD__
// There are several problematic interactions between FreeBSD's libthr and this
// jemalloc.
//
// 1. This malloc calls pthread_mutex_init at init, but in libthr this triggers
// an allocation, causing an infinite recursion.
// 2. Actually, this malloc assumes that lock initialization never triggers a
// memory allocation, even after initialization (see 'arena_new').
// 3. First use of a lock routine ('pthread_mutex_lock') in libthr triggers
// initialization of the process as a multi-threaded process. Unfortunately,
// libthr calls regular malloc as part of this bootstrap process.
//
// If there was no problem 3, we could have resolved this easily by using
// constant mutex initializers, since then libthr's uses its own internal
// allocator instead of regular malloc (this appears to have been the case for
// years now). However, problem 3 requires this malloc to provide some memory
// at places where it is not able to, so we need a way to divert standard
// allocator functions to some simple bootstrap allocator. And once we have
// done this, using constant mutex initializers looses most of its appeal,
// because allocations for problems 1 & 2 can be fulfilled by the simple
// allocator as well, without the drawback of being dependent on libthr's
// specific behavior.
//
// Since the init lock controls the 'malloc_initialized' flag, it is not
// possible to reliably check whether jemalloc is initialized in the case of
// multiple threads with the given tools (pthread cannot be used yet, but
// mutual exclusion is required). One solution would be to code simple
// user-space locks for this (e.g., spinlocks using GCC's builtins). But an
// even "simpler" solution is in fact to just remove the lock, on the ground
// that there must be some memory allocation before multithreading is enabled,
// so jemalloc is in fact always initialized before that point. And if there
// is not, we'll provoke it.
//
// At some point, I implemented a solution using __constructor__, as
// 'jemalloc_darwin_init', and tweaked the build so that it is included in
// executables (in platform/build/gecko_templates.mozbuild). But this was not
// enough: Clearly it could happen that some other library would be initialized
// before jemalloc, calling malloc in its contructor. Could have tried to work
// around this with constructor priorities, but this seemed fragile as well. So
// in the end, I kept the calls to 'malloc_init' from the interface's
// functions, and had to introduce 'malloc_initializing' to know when (part of
// the) calls should be diverted. I finally kept the constructor as well, just
// to absolutely guarantee that jemalloc is initialized during executable load,
// that is to say, before multi-threading happens, in case initialization in
// libthr or glib is removed at some point. It just doesn't call
// 'malloc_init_hard', contrary to Darwin's, but 'malloc_init' (because
// jemalloc normally has already been initialized at this point).
//
// During lock initialization, malloc is temporarily diverted to the bootstrap
// allocator to avoid harmful recursion. This is achieved using a flag
// indicating whether lock initialization is under way (in order to work also
// after malloc_init_hard() has completed). The flag *must* be per-thread,
// because creation of new arenas, which causes creation of new locks, can
// happen at unpredictable moments after multi-threading has been enabled (and
// malloc has been initialized), which means concurrent allocation requests can
// occur, and must not all be diverted. With this flag in place, and an
// additional change to ensure that libthr's multi-thread init is indeed done
// during mutex init (through 'pthread_lock_mutex'), there was no need to keep
// the 'malloc_initializing' flag (see previous paragraph).
//
// The most likely change this whole architecture is not immune to would be if
// jemalloc starts initializing new locks after malloc_init_hard() has finished
// but not under an existing lock (new arena's lock is currently initialized
// under the arenas lock), because bootstrap allocator functions are not
// thread-safe per se. If this happens, then a very simple spinlock
// implementation on top of GCC's atomics will be in order. But I don't think
// this is very likely to happen.
// Diverts key (de)allocation functions when jemalloc's mutexes are
// initializing (malloc_init_hard(), but also arena_new() and
// malloc_rtree_new(), as of this writing).
#define BA_DIVERT(code) \
do { \
if (in_mutex_init) { \
code; \
} \
} while (0)
// Bootstrap allocator
//
// It is not FreeBSD-specific, and could be used by any POSIX-compliant
// platform if needed.
//
// Allocates one page at a time (relies on 'pagesize' as defined above in this
// file), and returns memory from it. Does not accept allocations larger than a
// single page (minus alignment). Will waste space at end of pages. Never frees
// memory.
//
// All these constraints are not a problem, since this allocator is meant to
// serve only some requests at initialization (no more than a few kB).
// Number of really allocated bytes
static size_t ba_allocated_bn = 0;
// Number of requested bytes
static size_t ba_requested_bn = 0;
// Current address we are allocating from, or NULL if a new page has to be
// allocated.
static void *ba_cur_free = NULL;
static void ba_alloc_new_page()
{
ba_cur_free = mmap(NULL, pagesize, PROT_READ | PROT_WRITE,
MAP_ANON | MAP_PRIVATE, -1, 0);
if (ba_cur_free == MAP_FAILED)
abort();
ba_allocated_bn += pagesize;
}
// Returns the offset to point to have point a multiple of alignment
static size_t
ba_offset_to_aligned(uintptr_t point, size_t alignment) {
if (alignment != 0) {
size_t rest = point % alignment;
if (rest != 0)
return alignment - rest;
}
return 0;
}
static void * ba_memalign(size_t alignment, size_t size)
{
// We don't care about alignment being a power of 2, nor pagesize. Code
// below supports everything, provided that alignment divides the page
// size.
// Impose cache-line size minimum alignment, so that there is no cache
// trashing between fundamental structures.
if (alignment < CACHELINE)
alignment = CACHELINE;
if (size > pagesize ||
alignment > pagesize ||
size + alignment > pagesize ||
pagesize % alignment != 0)
abort();
// Address to be returned
uintptr_t cur_free;
// Allocate a new page if no current page (startup or previous one was
// exhausted) or there is not enough remaining space in it.
if (ba_cur_free == NULL) {
// No current page
ba_alloc_new_page();
cur_free = (uintptr_t)ba_cur_free;
} else {
cur_free = (uintptr_t)ba_cur_free;
uintptr_t off = cur_free % pagesize;
uintptr_t al_off = ba_offset_to_aligned(off, alignment);
if (off + al_off + size > pagesize) {
// Not enough room. Need a new page.
ba_alloc_new_page();
cur_free = (uintptr_t)ba_cur_free;
} else
// Account for alignment
cur_free += al_off;
}
// Compute the next free address
uintptr_t next_free = cur_free + size;
if (next_free % pagesize == 0 && size != 0)
next_free = 0;
// Set it
ba_cur_free = (void *)next_free;
// Stats
ba_requested_bn += size;
// Done
return (void *)cur_free;
}
static void * ba_malloc(size_t size)
{
// 64-bit alignment by default. ba_memalign imposes an even greater
// alignment anyway.
return ba_memalign(8, size);
}
static void * ba_calloc(size_t number, size_t size)
{
size_t const bn = number * size;
if ((bn < number || bn < size) && bn != 0)
// Overflow
abort();
void * const res = ba_malloc(bn);
memset(res, 0, bn);
return res;
}
static void ba_free(void * ptr) {
#ifdef MALLOC_DEBUG
malloc_printf("Bootstrap allocator: Request to free at %p\n", ptr);
#endif
// Do nothing
return;
}
#ifdef MALLOC_STATS
static void ba_print_stats() {
malloc_printf("Bootstrap allocator: %zu bytes requested, "
"%zu allocated\n",
ba_requested_bn, ba_allocated_bn);
}
#endif
__attribute__((constructor))
void
jemalloc_FreeBSD_init(void)
{
if (malloc_init())
abort();
}
#endif // #ifdef __FreeBSD__
#if !defined(MOZ_MEMORY_WINDOWS)
static
#endif
@ -5027,7 +5258,7 @@ malloc_init_hard(void)
malloc_zone_t* default_zone;
#endif
#ifndef MOZ_MEMORY_WINDOWS
#if !(defined(MOZ_MEMORY_WINDOWS) || defined(__FreeBSD__))
malloc_mutex_lock(&init_lock);
#endif
@ -5425,7 +5656,7 @@ MALLOC_OUT:
/* Allocate and initialize arenas. */
arenas = (arena_t **)base_alloc(sizeof(arena_t *) * narenas);
if (arenas == NULL) {
#ifndef MOZ_MEMORY_WINDOWS
#if !(defined(MOZ_MEMORY_WINDOWS) || defined(__FreeBSD__))
malloc_mutex_unlock(&init_lock);
#endif
return (true);
@ -5442,7 +5673,7 @@ MALLOC_OUT:
*/
arenas_extend(0);
if (arenas[0] == NULL) {
#ifndef MOZ_MEMORY_WINDOWS
#if !(defined(MOZ_MEMORY_WINDOWS) || defined(__FreeBSD__))
malloc_mutex_unlock(&init_lock);
#endif
return (true);
@ -5515,9 +5746,15 @@ MALLOC_OUT:
}
#endif
#ifndef MOZ_MEMORY_WINDOWS
#if defined(__FreeBSD__) && defined(MALLOC_STATS)
malloc_printf("Bootstrap allocator: malloc_init_hard stats:\n");
ba_print_stats();
#endif
#if !(defined(MOZ_MEMORY_WINDOWS) || defined(__FreeBSD__))
malloc_mutex_unlock(&init_lock);
#endif
return (false);
}
@ -5558,13 +5795,22 @@ malloc_shutdown()
#define DARWIN_ONLY(A)
#endif
#ifdef __FreeBSD__
#define FREEBSD_ONLY(code) code
#else
#define FREEBSD_ONLY(code)
#endif
MOZ_MEMORY_API void *
malloc_impl(size_t size)
{
void *ret;
DARWIN_ONLY(return (szone->malloc)(szone, size));
FREEBSD_ONLY(BA_DIVERT(return ba_malloc(size)));
void *ret;
if (malloc_init()) {
ret = NULL;
goto RETURN;
@ -5640,10 +5886,12 @@ MOZ_MEMORY_API
void *
MEMALIGN(size_t alignment, size_t size)
{
void *ret;
DARWIN_ONLY(return (szone->memalign)(szone, alignment, size));
FREEBSD_ONLY(BA_DIVERT(return ba_memalign(alignment, size)));
void *ret;
assert(((alignment - 1) & alignment) == 0);
if (malloc_init()) {
@ -5738,11 +5986,13 @@ valloc_impl(size_t size)
MOZ_MEMORY_API void *
calloc_impl(size_t num, size_t size)
{
DARWIN_ONLY(return (szone->calloc)(szone, num, size));
FREEBSD_ONLY(BA_DIVERT(return ba_calloc(num, size)));
void *ret;
size_t num_size;
DARWIN_ONLY(return (szone->calloc)(szone, num, size));
if (malloc_init()) {
num_size = 0;
ret = NULL;
@ -5857,10 +6107,12 @@ RETURN:
MOZ_MEMORY_API void
free_impl(void *ptr)
{
size_t offset;
DARWIN_ONLY((szone->free)(szone, ptr); return);
FREEBSD_ONLY(BA_DIVERT(return ba_free(ptr)));
size_t offset;
/*
* A version of idalloc that checks for NULL pointer but only for
* huge allocations assuming that CHUNK_ADDR2OFFSET(NULL) == 0.

View File

@ -38,7 +38,7 @@
#include <netinet/tcp_fsm.h> // for TCP connection states
#include <arpa/inet.h> // for inet_ntop()
#if __FreeBSD_version < 900000
#if !defined(__FreeBSD_version)
#include <utmp.h> // system users
#else
#include <utmpx.h>
@ -601,11 +601,7 @@ psutil_virtual_mem(PyObject *self, PyObject *args)
struct vmtotal vm;
int mib[] = {CTL_VM, VM_METER};
long pagesize = getpagesize();
#if __FreeBSD_version > 702101
long buffers;
#else
int buffers;
#endif
size_t buffers_size = sizeof(buffers);
if (sysctlbyname("vm.stats.vm.v_page_count", &total, &size, NULL, 0))
@ -724,13 +720,13 @@ psutil_cpu_times(PyObject *self, PyObject *args)
/*
* XXX
* These functions are available on FreeBSD 8 only.
* These functions were seen available on FreeBSD only.
* In the upper python layer we do various tricks to avoid crashing
* and/or to provide alternatives where possible.
*/
#if defined(__FreeBSD_version) && __FreeBSD_version >= 800000
#if defined(__FreeBSD_version)
/*
* Return files opened by process as a list of (path, fd) tuples.
* TODO: this is broken as it may report empty paths. 'procstat'
@ -943,11 +939,19 @@ psutil_sockaddr_matches(int family, int port, void *pcb_addr,
psutil_sockaddr_addrlen(family)) == 0);
}
#if __FreeBSD_version >= 1200026
static struct xtcpcb *
psutil_search_tcplist(char *buf, struct kinfo_file *kif)
{
struct xtcpcb *tp;
struct xinpcb *inp;
#else
static struct tcpcb *
psutil_search_tcplist(char *buf, struct kinfo_file *kif)
{
struct tcpcb *tp;
struct inpcb *inp;
#endif
struct xinpgen *xig, *oxig;
struct xsocket *so;
@ -955,9 +959,15 @@ psutil_search_tcplist(char *buf, struct kinfo_file *kif)
for (xig = (struct xinpgen *)((char *)xig + xig->xig_len);
xig->xig_len > sizeof(struct xinpgen);
xig = (struct xinpgen *)((char *)xig + xig->xig_len)) {
#if __FreeBSD_version >= 1200026
tp = (struct xtcpcb *)xig;
inp = &tp->xt_inp;
so = &inp->xi_socket;
#else
tp = &((struct xtcpcb *)xig)->xt_tp;
inp = &((struct xtcpcb *)xig)->xt_inp;
so = &((struct xtcpcb *)xig)->xt_socket;
#endif
if (so->so_type != kif->kf_sock_type ||
so->xso_family != kif->kf_sock_domain ||
@ -967,20 +977,36 @@ psutil_search_tcplist(char *buf, struct kinfo_file *kif)
if (kif->kf_sock_domain == AF_INET) {
if (!psutil_sockaddr_matches(
AF_INET, inp->inp_lport, &inp->inp_laddr,
#if __FreeBSD_version < 1200031
&kif->kf_sa_local))
#else
&kif->kf_un.kf_sock.kf_sa_local))
#endif
continue;
if (!psutil_sockaddr_matches(
AF_INET, inp->inp_fport, &inp->inp_faddr,
#if __FreeBSD_version < 1200031
&kif->kf_sa_peer))
#else
&kif->kf_un.kf_sock.kf_sa_peer))
#endif
continue;
} else {
if (!psutil_sockaddr_matches(
AF_INET6, inp->inp_lport, &inp->in6p_laddr,
#if __FreeBSD_version < 1200031
&kif->kf_sa_local))
#else
&kif->kf_un.kf_sock.kf_sa_local))
#endif
continue;
if (!psutil_sockaddr_matches(
AF_INET6, inp->inp_fport, &inp->in6p_faddr,
#if __FreeBSD_version < 1200031
&kif->kf_sa_peer))
#else
&kif->kf_un.kf_sock.kf_sa_peer))
#endif
continue;
}
@ -1005,7 +1031,11 @@ psutil_proc_connections(PyObject *self, PyObject *args)
struct kinfo_file *freep = NULL;
struct kinfo_file *kif;
char *tcplist = NULL;
#if __FreeBSD_version >= 1200026
struct xtcpcb *tcp;
#else
struct tcpcb *tcp;
#endif
PyObject *retList = PyList_New(0);
PyObject *tuple = NULL;
@ -1074,19 +1104,35 @@ psutil_proc_connections(PyObject *self, PyObject *args)
inet_ntop(
kif->kf_sock_domain,
psutil_sockaddr_addr(kif->kf_sock_domain,
#if __FreeBSD_version < 1200031
&kif->kf_sa_local),
#else
&kif->kf_un.kf_sock.kf_sa_local),
#endif
lip,
sizeof(lip));
inet_ntop(
kif->kf_sock_domain,
psutil_sockaddr_addr(kif->kf_sock_domain,
#if __FreeBSD_version < 1200031
&kif->kf_sa_peer),
#else
&kif->kf_un.kf_sock.kf_sa_peer),
#endif
rip,
sizeof(rip));
lport = htons(psutil_sockaddr_port(kif->kf_sock_domain,
#if __FreeBSD_version < 1200031
&kif->kf_sa_local));
#else
&kif->kf_un.kf_sock.kf_sa_local));
#endif
rport = htons(psutil_sockaddr_port(kif->kf_sock_domain,
#if __FreeBSD_version < 1200031
&kif->kf_sa_peer));
#else
&kif->kf_un.kf_sock.kf_sa_peer));
#endif
// construct python tuple/list
laddr = Py_BuildValue("(si)", lip, lport);
@ -1115,7 +1161,11 @@ psutil_proc_connections(PyObject *self, PyObject *args)
else if (kif->kf_sock_domain == AF_UNIX) {
struct sockaddr_un *sun;
#if __FreeBSD_version < 1200031
sun = (struct sockaddr_un *)&kif->kf_sa_local;
#else
sun = (struct sockaddr_un *)&kif->kf_un.kf_sock.kf_sa_local;
#endif
snprintf(
path, sizeof(path), "%.*s",
(int)(sun->sun_len - (sizeof(*sun) - sizeof(sun->sun_path))),
@ -1626,7 +1676,7 @@ psutil_users(PyObject *self, PyObject *args)
if (ret_list == NULL)
return NULL;
#if __FreeBSD_version < 900000
#if !defined(__FreeBSD_version)
struct utmp ut;
FILE *fp;
@ -1736,7 +1786,7 @@ psutil_get_pid_from_sock(int sock_hash)
struct xfile *xf;
int hash, n;
for (xf = psutil_xfiles, n = 0; n < psutil_nxfiles; ++n, ++xf) {
if (xf->xf_data == NULL)
if (xf->xf_data == 0)
continue;
hash = (int)((uintptr_t)xf->xf_data % HASHSIZE);
if (sock_hash == hash)
@ -1754,7 +1804,11 @@ int psutil_gather_inet(int proto, PyObject *py_retlist)
struct xinpgen *xig, *exig;
struct xinpcb *xip;
struct xtcpcb *xtp;
#if defined(__FreeBSD_version) && __FreeBSD_version >= 1200026
struct xinpcb *inp;
#else
struct inpcb *inp;
#endif
struct xsocket *so;
const char *varname = NULL;
size_t len, bufsize;
@ -1820,8 +1874,13 @@ int psutil_gather_inet(int proto, PyObject *py_retlist)
goto error;
}
inp = &xtp->xt_inp;
#if defined(__FreeBSD_version) && __FreeBSD_version >= 1200026
so = &inp->xi_socket;
status = xtp->t_state;
#else
so = &xtp->xt_socket;
status = xtp->xt_tp.t_state;
#endif
break;
case IPPROTO_UDP:
xip = (struct xinpcb *)xig;
@ -1830,7 +1889,11 @@ int psutil_gather_inet(int proto, PyObject *py_retlist)
"struct xinpcb size mismatch");
goto error;
}
#if defined(__FreeBSD_version) && __FreeBSD_version >= 1200026
inp = xip;
#else
inp = &xip->xi_inp;
#endif
so = &xip->xi_socket;
status = PSUTIL_CONN_NONE;
break;
@ -2166,7 +2229,7 @@ PsutilMethods[] =
"Return process CPU affinity."},
{"proc_cpu_affinity_set", psutil_proc_cpu_affinity_set, METH_VARARGS,
"Set process CPU affinity."},
#if defined(__FreeBSD_version) && __FreeBSD_version >= 800000
#if defined(__FreeBSD_version)
{"proc_open_files", psutil_proc_open_files, METH_VARARGS,
"Return files opened by process as a list of (path, fd) tuples"},
{"proc_cwd", psutil_proc_cwd, METH_VARARGS,
@ -2191,7 +2254,7 @@ PsutilMethods[] =
"Return swap mem stats"},
{"cpu_times", psutil_cpu_times, METH_VARARGS,
"Return system cpu times as a tuple (user, system, nice, idle, irc)"},
#if defined(__FreeBSD_version) && __FreeBSD_version >= 800000
#if defined(__FreeBSD_version)
{"per_cpu_times", psutil_per_cpu_times, METH_VARARGS,
"Return system per-cpu times as a list of tuples"},
#endif

View File

@ -29,7 +29,7 @@ static PyObject* psutil_proc_uids(PyObject* self, PyObject* args);
static PyObject* psutil_proc_cpu_affinity_get(PyObject* self, PyObject* args);
static PyObject* psutil_proc_cpu_affinity_set(PyObject* self, PyObject* args);
#if defined(__FreeBSD_version) && __FreeBSD_version >= 800000
#if defined(__FreeBSD_version)
static PyObject* psutil_proc_open_files(PyObject* self, PyObject* args);
static PyObject* psutil_proc_cwd(PyObject* self, PyObject* args);
#endif
@ -48,6 +48,6 @@ static PyObject* psutil_swap_mem(PyObject* self, PyObject* args);
static PyObject* psutil_users(PyObject* self, PyObject* args);
static PyObject* psutil_virtual_mem(PyObject* self, PyObject* args);
#if defined(__FreeBSD_version) && __FreeBSD_version >= 800000
#if defined(__FreeBSD_version)
static PyObject* psutil_per_cpu_times(PyObject* self, PyObject* args);
#endif