version 1.1.2.7, 2013/06/25 08:45:50
|
version 1.1.2.10, 2013/06/25 09:29:34
|
Line 163 io_etherOpen(const char *csIface, int flags, int whdr,
|
Line 163 io_etherOpen(const char *csIface, int flags, int whdr,
|
io_etherClose(eth, zcbuf); |
io_etherClose(eth, zcbuf); |
return -1; |
return -1; |
} |
} |
|
memset(&ifr, 0, sizeof ifr); |
strlcpy(ifr.ifr_name, szStr, sizeof ifr.ifr_name); |
strlcpy(ifr.ifr_name, szStr, sizeof ifr.ifr_name); |
if (ioctl(eth, BIOCSETIF, &ifr) == -1) { |
if (ioctl(eth, BIOCSETIF, &ifr) == -1) { |
LOGERR; |
LOGERR; |
Line 198 io_etherSend(int eth, const void *buf, size_t buflen)
|
Line 199 io_etherSend(int eth, const void *buf, size_t buflen)
|
} |
} |
|
|
#ifdef __FreeBSD__ |
#ifdef __FreeBSD__ |
|
static inline void |
|
ackZCbuf(struct bpf_zbuf_header * __restrict bzh) |
|
{ |
|
atomic_store_rel_int(&bzh->bzh_user_gen, bzh->bzh_kernel_gen); |
|
} |
|
|
|
static inline int |
|
chkZCbuf(struct bpf_zbuf_header * __restrict bzh) |
|
{ |
|
/* return true if userspace owns buffer, and false otherwise. */ |
|
return (bzh->bzh_user_gen != atomic_load_acq_int(&bzh->bzh_kernel_gen)); |
|
} |
|
|
static inline ssize_t |
static inline ssize_t |
nextZCbuf(void ** __restrict zcache, struct bpf_zbuf * __restrict zbuf, |
nextZCbuf(void ** __restrict zcache, struct bpf_zbuf * __restrict zbuf, |
const void * __restrict buf) |
const void * __restrict buf) |
Line 207 nextZCbuf(void ** __restrict zcache, struct bpf_zbuf *
|
Line 221 nextZCbuf(void ** __restrict zcache, struct bpf_zbuf *
|
|
|
if (!*zcache || *zcache == zbuf->bz_bufb) { |
if (!*zcache || *zcache == zbuf->bz_bufb) { |
bzh = (struct bpf_zbuf_header *) zbuf->bz_bufa; |
bzh = (struct bpf_zbuf_header *) zbuf->bz_bufa; |
if (bzh->bzh_user_gen != atomic_load_acq_int(&bzh->bzh_kernel_gen)) { | if (chkZCbuf(bzh)) { |
rlen = atomic_load_acq_int(&bzh->bzh_kernel_len); |
rlen = atomic_load_acq_int(&bzh->bzh_kernel_len); |
*zcache = zbuf->bz_bufa; |
*zcache = zbuf->bz_bufa; |
if (buf) |
if (buf) |
buf = ((caddr_t) *zcache) + sizeof(struct bpf_zbuf_header); |
buf = ((caddr_t) *zcache) + sizeof(struct bpf_zbuf_header); |
} | ackZCbuf(bzh); |
| } else |
| io_SetErr(EAGAIN, "kernel owns the buffer"); |
} else if (*zcache == zbuf->bz_bufa) { |
} else if (*zcache == zbuf->bz_bufa) { |
bzh = (struct bpf_zbuf_header *) zbuf->bz_bufb; |
bzh = (struct bpf_zbuf_header *) zbuf->bz_bufb; |
if (bzh->bzh_user_gen != atomic_load_acq_int(&bzh->bzh_kernel_gen)) { | if (chkZCbuf(bzh)) { |
rlen = atomic_load_acq_int(&bzh->bzh_kernel_len); |
rlen = atomic_load_acq_int(&bzh->bzh_kernel_len); |
*zcache = zbuf->bz_bufb; |
*zcache = zbuf->bz_bufb; |
if (buf) |
if (buf) |
buf = ((caddr_t) *zcache) + sizeof(struct bpf_zbuf_header); |
buf = ((caddr_t) *zcache) + sizeof(struct bpf_zbuf_header); |
} | ackZCbuf(bzh); |
| } else |
| io_SetErr(EAGAIN, "kernel owns the buffer"); |
} |
} |
|
|
return rlen; |
return rlen; |