libev_scheduler 0.1
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +7 -0
- data/.github/test.yml +31 -0
- data/.gitignore +58 -0
- data/CHANGELOG.md +0 -0
- data/Gemfile +3 -0
- data/Gemfile.lock +32 -0
- data/LICENSE +21 -0
- data/README.md +5 -0
- data/Rakefile +18 -0
- data/TODO.md +0 -0
- data/examples/io.rb +31 -0
- data/examples/sleep.rb +10 -0
- data/ext/libev/Changes +548 -0
- data/ext/libev/LICENSE +37 -0
- data/ext/libev/README +59 -0
- data/ext/libev/README.embed +3 -0
- data/ext/libev/ev.c +5279 -0
- data/ext/libev/ev.h +856 -0
- data/ext/libev/ev_epoll.c +296 -0
- data/ext/libev/ev_kqueue.c +224 -0
- data/ext/libev/ev_linuxaio.c +642 -0
- data/ext/libev/ev_poll.c +156 -0
- data/ext/libev/ev_port.c +192 -0
- data/ext/libev/ev_select.c +316 -0
- data/ext/libev/ev_vars.h +215 -0
- data/ext/libev/ev_win32.c +162 -0
- data/ext/libev/ev_wrap.h +216 -0
- data/ext/libev/test_libev_win32.c +123 -0
- data/ext/libev_scheduler/extconf.rb +22 -0
- data/ext/libev_scheduler/libev.c +2 -0
- data/ext/libev_scheduler/libev.h +11 -0
- data/ext/libev_scheduler/libev_scheduler_ext.c +5 -0
- data/ext/libev_scheduler/scheduler.c +330 -0
- data/lib/libev_scheduler.rb +23 -0
- data/lib/libev_scheduler/version.rb +5 -0
- data/libev_scheduler.gemspec +26 -0
- data/test/run.rb +5 -0
- data/test/test_enumerator.rb +45 -0
- data/test/test_io.rb +57 -0
- data/test/test_mutex.rb +230 -0
- data/test/test_process.rb +38 -0
- data/test/test_sleep.rb +52 -0
- metadata +134 -0
@@ -0,0 +1,296 @@
|
|
1
|
+
/*
|
2
|
+
* libev epoll fd activity backend
|
3
|
+
*
|
4
|
+
* Copyright (c) 2007,2008,2009,2010,2011,2016,2017,2019 Marc Alexander Lehmann <libev@schmorp.de>
|
5
|
+
* All rights reserved.
|
6
|
+
*
|
7
|
+
* Redistribution and use in source and binary forms, with or without modifica-
|
8
|
+
* tion, are permitted provided that the following conditions are met:
|
9
|
+
*
|
10
|
+
* 1. Redistributions of source code must retain the above copyright notice,
|
11
|
+
* this list of conditions and the following disclaimer.
|
12
|
+
*
|
13
|
+
* 2. Redistributions in binary form must reproduce the above copyright
|
14
|
+
* notice, this list of conditions and the following disclaimer in the
|
15
|
+
* documentation and/or other materials provided with the distribution.
|
16
|
+
*
|
17
|
+
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR IMPLIED
|
18
|
+
* WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MER-
|
19
|
+
* CHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO
|
20
|
+
* EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPE-
|
21
|
+
* CIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
|
22
|
+
* PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS;
|
23
|
+
* OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY,
|
24
|
+
* WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTH-
|
25
|
+
* ERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED
|
26
|
+
* OF THE POSSIBILITY OF SUCH DAMAGE.
|
27
|
+
*
|
28
|
+
* Alternatively, the contents of this file may be used under the terms of
|
29
|
+
* the GNU General Public License ("GPL") version 2 or any later version,
|
30
|
+
* in which case the provisions of the GPL are applicable instead of
|
31
|
+
* the above. If you wish to allow the use of your version of this file
|
32
|
+
* only under the terms of the GPL and not to allow others to use your
|
33
|
+
* version of this file under the BSD license, indicate your decision
|
34
|
+
* by deleting the provisions above and replace them with the notice
|
35
|
+
* and other provisions required by the GPL. If you do not delete the
|
36
|
+
* provisions above, a recipient may use your version of this file under
|
37
|
+
* either the BSD or the GPL.
|
38
|
+
*/
|
39
|
+
|
40
|
+
/*
|
41
|
+
* general notes about epoll:
|
42
|
+
*
|
43
|
+
* a) epoll silently removes fds from the fd set. as nothing tells us
|
44
|
+
* that an fd has been removed otherwise, we have to continually
|
45
|
+
* "rearm" fds that we suspect *might* have changed (same
|
46
|
+
* problem with kqueue, but much less costly there).
|
47
|
+
* b) the fact that ADD != MOD creates a lot of extra syscalls due to a)
|
48
|
+
* and seems not to have any advantage.
|
49
|
+
* c) the inability to handle fork or file descriptors (think dup)
|
50
|
+
* limits the applicability over poll, so this is not a generic
|
51
|
+
* poll replacement.
|
52
|
+
* d) epoll doesn't work the same as select with many file descriptors
|
53
|
+
* (such as files). while not critical, no other advanced interface
|
54
|
+
* seems to share this (rather non-unixy) limitation.
|
55
|
+
* e) epoll claims to be embeddable, but in practise you never get
|
56
|
+
* a ready event for the epoll fd (broken: <=2.6.26, working: >=2.6.32).
|
57
|
+
* f) epoll_ctl returning EPERM means the fd is always ready.
|
58
|
+
*
|
59
|
+
* lots of "weird code" and complication handling in this file is due
|
60
|
+
* to these design problems with epoll, as we try very hard to avoid
|
61
|
+
* epoll_ctl syscalls for common usage patterns and handle the breakage
|
62
|
+
* ensuing from receiving events for closed and otherwise long gone
|
63
|
+
* file descriptors.
|
64
|
+
*/
|
65
|
+
|
66
|
+
#include <sys/epoll.h>
|
67
|
+
|
68
|
+
#define EV_EMASK_EPERM 0x80
|
69
|
+
|
70
|
+
static void
|
71
|
+
epoll_modify (EV_P_ int fd, int oev, int nev)
|
72
|
+
{
|
73
|
+
struct epoll_event ev;
|
74
|
+
unsigned char oldmask;
|
75
|
+
|
76
|
+
/*
|
77
|
+
* we handle EPOLL_CTL_DEL by ignoring it here
|
78
|
+
* on the assumption that the fd is gone anyways
|
79
|
+
* if that is wrong, we have to handle the spurious
|
80
|
+
* event in epoll_poll.
|
81
|
+
* if the fd is added again, we try to ADD it, and, if that
|
82
|
+
* fails, we assume it still has the same eventmask.
|
83
|
+
*/
|
84
|
+
if (!nev)
|
85
|
+
return;
|
86
|
+
|
87
|
+
oldmask = anfds [fd].emask;
|
88
|
+
anfds [fd].emask = nev;
|
89
|
+
|
90
|
+
/* store the generation counter in the upper 32 bits, the fd in the lower 32 bits */
|
91
|
+
ev.data.u64 = (uint64_t)(uint32_t)fd
|
92
|
+
| ((uint64_t)(uint32_t)++anfds [fd].egen << 32);
|
93
|
+
ev.events = (nev & EV_READ ? EPOLLIN : 0)
|
94
|
+
| (nev & EV_WRITE ? EPOLLOUT : 0);
|
95
|
+
|
96
|
+
if (expect_true (!epoll_ctl (backend_fd, oev && oldmask != nev ? EPOLL_CTL_MOD : EPOLL_CTL_ADD, fd, &ev)))
|
97
|
+
return;
|
98
|
+
|
99
|
+
if (expect_true (errno == ENOENT))
|
100
|
+
{
|
101
|
+
/* if ENOENT then the fd went away, so try to do the right thing */
|
102
|
+
if (!nev)
|
103
|
+
goto dec_egen;
|
104
|
+
|
105
|
+
if (!epoll_ctl (backend_fd, EPOLL_CTL_ADD, fd, &ev))
|
106
|
+
return;
|
107
|
+
}
|
108
|
+
else if (expect_true (errno == EEXIST))
|
109
|
+
{
|
110
|
+
/* EEXIST means we ignored a previous DEL, but the fd is still active */
|
111
|
+
/* if the kernel mask is the same as the new mask, we assume it hasn't changed */
|
112
|
+
if (oldmask == nev)
|
113
|
+
goto dec_egen;
|
114
|
+
|
115
|
+
if (!epoll_ctl (backend_fd, EPOLL_CTL_MOD, fd, &ev))
|
116
|
+
return;
|
117
|
+
}
|
118
|
+
else if (expect_true (errno == EPERM))
|
119
|
+
{
|
120
|
+
/* EPERM means the fd is always ready, but epoll is too snobbish */
|
121
|
+
/* to handle it, unlike select or poll. */
|
122
|
+
anfds [fd].emask = EV_EMASK_EPERM;
|
123
|
+
|
124
|
+
/* add fd to epoll_eperms, if not already inside */
|
125
|
+
if (!(oldmask & EV_EMASK_EPERM))
|
126
|
+
{
|
127
|
+
array_needsize (int, epoll_eperms, epoll_epermmax, epoll_epermcnt + 1, array_needsize_noinit);
|
128
|
+
epoll_eperms [epoll_epermcnt++] = fd;
|
129
|
+
}
|
130
|
+
|
131
|
+
return;
|
132
|
+
}
|
133
|
+
else
|
134
|
+
assert (("libev: I/O watcher with invalid fd found in epoll_ctl", errno != EBADF && errno != ELOOP && errno != EINVAL));
|
135
|
+
|
136
|
+
fd_kill (EV_A_ fd);
|
137
|
+
|
138
|
+
dec_egen:
|
139
|
+
/* we didn't successfully call epoll_ctl, so decrement the generation counter again */
|
140
|
+
--anfds [fd].egen;
|
141
|
+
}
|
142
|
+
|
143
|
+
static void
|
144
|
+
epoll_poll (EV_P_ ev_tstamp timeout)
|
145
|
+
{
|
146
|
+
int i;
|
147
|
+
int eventcnt;
|
148
|
+
|
149
|
+
if (expect_false (epoll_epermcnt))
|
150
|
+
timeout = 0.;
|
151
|
+
|
152
|
+
/* epoll wait times cannot be larger than (LONG_MAX - 999UL) / HZ msecs, which is below */
|
153
|
+
/* the default libev max wait time, however. */
|
154
|
+
EV_RELEASE_CB;
|
155
|
+
eventcnt = epoll_wait (backend_fd, epoll_events, epoll_eventmax, timeout * 1e3);
|
156
|
+
EV_ACQUIRE_CB;
|
157
|
+
|
158
|
+
if (expect_false (eventcnt < 0))
|
159
|
+
{
|
160
|
+
if (errno != EINTR)
|
161
|
+
ev_syserr ("(libev) epoll_wait");
|
162
|
+
|
163
|
+
return;
|
164
|
+
}
|
165
|
+
|
166
|
+
for (i = 0; i < eventcnt; ++i)
|
167
|
+
{
|
168
|
+
struct epoll_event *ev = epoll_events + i;
|
169
|
+
|
170
|
+
int fd = (uint32_t)ev->data.u64; /* mask out the lower 32 bits */
|
171
|
+
int want = anfds [fd].events;
|
172
|
+
int got = (ev->events & (EPOLLOUT | EPOLLERR | EPOLLHUP) ? EV_WRITE : 0)
|
173
|
+
| (ev->events & (EPOLLIN | EPOLLERR | EPOLLHUP) ? EV_READ : 0);
|
174
|
+
|
175
|
+
/*
|
176
|
+
* check for spurious notification.
|
177
|
+
* this only finds spurious notifications on egen updates
|
178
|
+
* other spurious notifications will be found by epoll_ctl, below
|
179
|
+
* we assume that fd is always in range, as we never shrink the anfds array
|
180
|
+
*/
|
181
|
+
if (expect_false ((uint32_t)anfds [fd].egen != (uint32_t)(ev->data.u64 >> 32)))
|
182
|
+
{
|
183
|
+
/* recreate kernel state */
|
184
|
+
postfork |= 2;
|
185
|
+
continue;
|
186
|
+
}
|
187
|
+
|
188
|
+
if (expect_false (got & ~want))
|
189
|
+
{
|
190
|
+
anfds [fd].emask = want;
|
191
|
+
|
192
|
+
/*
|
193
|
+
* we received an event but are not interested in it, try mod or del
|
194
|
+
* this often happens because we optimistically do not unregister fds
|
195
|
+
* when we are no longer interested in them, but also when we get spurious
|
196
|
+
* notifications for fds from another process. this is partially handled
|
197
|
+
* above with the gencounter check (== our fd is not the event fd), and
|
198
|
+
* partially here, when epoll_ctl returns an error (== a child has the fd
|
199
|
+
* but we closed it).
|
200
|
+
*/
|
201
|
+
ev->events = (want & EV_READ ? EPOLLIN : 0)
|
202
|
+
| (want & EV_WRITE ? EPOLLOUT : 0);
|
203
|
+
|
204
|
+
/* pre-2.6.9 kernels require a non-null pointer with EPOLL_CTL_DEL, */
|
205
|
+
/* which is fortunately easy to do for us. */
|
206
|
+
if (epoll_ctl (backend_fd, want ? EPOLL_CTL_MOD : EPOLL_CTL_DEL, fd, ev))
|
207
|
+
{
|
208
|
+
postfork |= 2; /* an error occurred, recreate kernel state */
|
209
|
+
continue;
|
210
|
+
}
|
211
|
+
}
|
212
|
+
|
213
|
+
fd_event (EV_A_ fd, got);
|
214
|
+
}
|
215
|
+
|
216
|
+
/* if the receive array was full, increase its size */
|
217
|
+
if (expect_false (eventcnt == epoll_eventmax))
|
218
|
+
{
|
219
|
+
ev_free (epoll_events);
|
220
|
+
epoll_eventmax = array_nextsize (sizeof (struct epoll_event), epoll_eventmax, epoll_eventmax + 1);
|
221
|
+
epoll_events = (struct epoll_event *)ev_malloc (sizeof (struct epoll_event) * epoll_eventmax);
|
222
|
+
}
|
223
|
+
|
224
|
+
/* now synthesize events for all fds where epoll fails, while select works... */
|
225
|
+
for (i = epoll_epermcnt; i--; )
|
226
|
+
{
|
227
|
+
int fd = epoll_eperms [i];
|
228
|
+
unsigned char events = anfds [fd].events & (EV_READ | EV_WRITE);
|
229
|
+
|
230
|
+
if (anfds [fd].emask & EV_EMASK_EPERM && events)
|
231
|
+
fd_event (EV_A_ fd, events);
|
232
|
+
else
|
233
|
+
{
|
234
|
+
epoll_eperms [i] = epoll_eperms [--epoll_epermcnt];
|
235
|
+
anfds [fd].emask = 0;
|
236
|
+
}
|
237
|
+
}
|
238
|
+
}
|
239
|
+
|
240
|
+
static int
|
241
|
+
epoll_epoll_create (void)
|
242
|
+
{
|
243
|
+
int fd;
|
244
|
+
|
245
|
+
#if defined EPOLL_CLOEXEC && !defined __ANDROID__
|
246
|
+
fd = epoll_create1 (EPOLL_CLOEXEC);
|
247
|
+
|
248
|
+
if (fd < 0 && (errno == EINVAL || errno == ENOSYS))
|
249
|
+
#endif
|
250
|
+
{
|
251
|
+
fd = epoll_create (256);
|
252
|
+
|
253
|
+
if (fd >= 0)
|
254
|
+
fcntl (fd, F_SETFD, FD_CLOEXEC);
|
255
|
+
}
|
256
|
+
|
257
|
+
return fd;
|
258
|
+
}
|
259
|
+
|
260
|
+
inline_size
|
261
|
+
int
|
262
|
+
epoll_init (EV_P_ int flags)
|
263
|
+
{
|
264
|
+
if ((backend_fd = epoll_epoll_create ()) < 0)
|
265
|
+
return 0;
|
266
|
+
|
267
|
+
backend_mintime = 1e-3; /* epoll does sometimes return early, this is just to avoid the worst */
|
268
|
+
backend_modify = epoll_modify;
|
269
|
+
backend_poll = epoll_poll;
|
270
|
+
|
271
|
+
epoll_eventmax = 64; /* initial number of events receivable per poll */
|
272
|
+
epoll_events = (struct epoll_event *)ev_malloc (sizeof (struct epoll_event) * epoll_eventmax);
|
273
|
+
|
274
|
+
return EVBACKEND_EPOLL;
|
275
|
+
}
|
276
|
+
|
277
|
+
inline_size
|
278
|
+
void
|
279
|
+
epoll_destroy (EV_P)
|
280
|
+
{
|
281
|
+
ev_free (epoll_events);
|
282
|
+
array_free (epoll_eperm, EMPTY);
|
283
|
+
}
|
284
|
+
|
285
|
+
inline_size
|
286
|
+
void
|
287
|
+
epoll_fork (EV_P)
|
288
|
+
{
|
289
|
+
close (backend_fd);
|
290
|
+
|
291
|
+
while ((backend_fd = epoll_epoll_create ()) < 0)
|
292
|
+
ev_syserr ("(libev) epoll_create");
|
293
|
+
|
294
|
+
fd_rearm_all (EV_A);
|
295
|
+
}
|
296
|
+
|
@@ -0,0 +1,224 @@
|
|
1
|
+
/*
|
2
|
+
* libev kqueue backend
|
3
|
+
*
|
4
|
+
* Copyright (c) 2007,2008,2009,2010,2011,2012,2013,2016,2019 Marc Alexander Lehmann <libev@schmorp.de>
|
5
|
+
* All rights reserved.
|
6
|
+
*
|
7
|
+
* Redistribution and use in source and binary forms, with or without modifica-
|
8
|
+
* tion, are permitted provided that the following conditions are met:
|
9
|
+
*
|
10
|
+
* 1. Redistributions of source code must retain the above copyright notice,
|
11
|
+
* this list of conditions and the following disclaimer.
|
12
|
+
*
|
13
|
+
* 2. Redistributions in binary form must reproduce the above copyright
|
14
|
+
* notice, this list of conditions and the following disclaimer in the
|
15
|
+
* documentation and/or other materials provided with the distribution.
|
16
|
+
*
|
17
|
+
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR IMPLIED
|
18
|
+
* WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MER-
|
19
|
+
* CHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO
|
20
|
+
* EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPE-
|
21
|
+
* CIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
|
22
|
+
* PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS;
|
23
|
+
* OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY,
|
24
|
+
* WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTH-
|
25
|
+
* ERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED
|
26
|
+
* OF THE POSSIBILITY OF SUCH DAMAGE.
|
27
|
+
*
|
28
|
+
* Alternatively, the contents of this file may be used under the terms of
|
29
|
+
* the GNU General Public License ("GPL") version 2 or any later version,
|
30
|
+
* in which case the provisions of the GPL are applicable instead of
|
31
|
+
* the above. If you wish to allow the use of your version of this file
|
32
|
+
* only under the terms of the GPL and not to allow others to use your
|
33
|
+
* version of this file under the BSD license, indicate your decision
|
34
|
+
* by deleting the provisions above and replace them with the notice
|
35
|
+
* and other provisions required by the GPL. If you do not delete the
|
36
|
+
* provisions above, a recipient may use your version of this file under
|
37
|
+
* either the BSD or the GPL.
|
38
|
+
*/
|
39
|
+
|
40
|
+
#include <sys/types.h>
|
41
|
+
#include <sys/time.h>
|
42
|
+
#include <sys/event.h>
|
43
|
+
#include <string.h>
|
44
|
+
#include <errno.h>
|
45
|
+
|
46
|
+
inline_speed
|
47
|
+
void
|
48
|
+
kqueue_change (EV_P_ int fd, int filter, int flags, int fflags)
|
49
|
+
{
|
50
|
+
++kqueue_changecnt;
|
51
|
+
array_needsize (struct kevent, kqueue_changes, kqueue_changemax, kqueue_changecnt, array_needsize_noinit);
|
52
|
+
|
53
|
+
EV_SET (&kqueue_changes [kqueue_changecnt - 1], fd, filter, flags, fflags, 0, 0);
|
54
|
+
}
|
55
|
+
|
56
|
+
/* OS X at least needs this */
|
57
|
+
#ifndef EV_ENABLE
|
58
|
+
# define EV_ENABLE 0
|
59
|
+
#endif
|
60
|
+
#ifndef NOTE_EOF
|
61
|
+
# define NOTE_EOF 0
|
62
|
+
#endif
|
63
|
+
|
64
|
+
static void
|
65
|
+
kqueue_modify (EV_P_ int fd, int oev, int nev)
|
66
|
+
{
|
67
|
+
if (oev != nev)
|
68
|
+
{
|
69
|
+
if (oev & EV_READ)
|
70
|
+
kqueue_change (EV_A_ fd, EVFILT_READ , EV_DELETE, 0);
|
71
|
+
|
72
|
+
if (oev & EV_WRITE)
|
73
|
+
kqueue_change (EV_A_ fd, EVFILT_WRITE, EV_DELETE, 0);
|
74
|
+
}
|
75
|
+
|
76
|
+
/* to detect close/reopen reliably, we have to re-add */
|
77
|
+
/* event requests even when oev == nev */
|
78
|
+
|
79
|
+
if (nev & EV_READ)
|
80
|
+
kqueue_change (EV_A_ fd, EVFILT_READ , EV_ADD | EV_ENABLE, NOTE_EOF);
|
81
|
+
|
82
|
+
if (nev & EV_WRITE)
|
83
|
+
kqueue_change (EV_A_ fd, EVFILT_WRITE, EV_ADD | EV_ENABLE, NOTE_EOF);
|
84
|
+
}
|
85
|
+
|
86
|
+
static void
|
87
|
+
kqueue_poll (EV_P_ ev_tstamp timeout)
|
88
|
+
{
|
89
|
+
int res, i;
|
90
|
+
struct timespec ts;
|
91
|
+
|
92
|
+
/* need to resize so there is enough space for errors */
|
93
|
+
if (kqueue_changecnt > kqueue_eventmax)
|
94
|
+
{
|
95
|
+
ev_free (kqueue_events);
|
96
|
+
kqueue_eventmax = array_nextsize (sizeof (struct kevent), kqueue_eventmax, kqueue_changecnt);
|
97
|
+
kqueue_events = (struct kevent *)ev_malloc (sizeof (struct kevent) * kqueue_eventmax);
|
98
|
+
}
|
99
|
+
|
100
|
+
EV_RELEASE_CB;
|
101
|
+
EV_TS_SET (ts, timeout);
|
102
|
+
res = kevent (backend_fd, kqueue_changes, kqueue_changecnt, kqueue_events, kqueue_eventmax, &ts);
|
103
|
+
EV_ACQUIRE_CB;
|
104
|
+
kqueue_changecnt = 0;
|
105
|
+
|
106
|
+
if (expect_false (res < 0))
|
107
|
+
{
|
108
|
+
if (errno != EINTR)
|
109
|
+
ev_syserr ("(libev) kqueue kevent");
|
110
|
+
|
111
|
+
return;
|
112
|
+
}
|
113
|
+
|
114
|
+
for (i = 0; i < res; ++i)
|
115
|
+
{
|
116
|
+
int fd = kqueue_events [i].ident;
|
117
|
+
|
118
|
+
if (expect_false (kqueue_events [i].flags & EV_ERROR))
|
119
|
+
{
|
120
|
+
int err = kqueue_events [i].data;
|
121
|
+
|
122
|
+
/* we are only interested in errors for fds that we are interested in :) */
|
123
|
+
if (anfds [fd].events)
|
124
|
+
{
|
125
|
+
if (err == ENOENT) /* resubmit changes on ENOENT */
|
126
|
+
kqueue_modify (EV_A_ fd, 0, anfds [fd].events);
|
127
|
+
else if (err == EBADF) /* on EBADF, we re-check the fd */
|
128
|
+
{
|
129
|
+
if (fd_valid (fd))
|
130
|
+
kqueue_modify (EV_A_ fd, 0, anfds [fd].events);
|
131
|
+
else
|
132
|
+
{
|
133
|
+
assert (("libev: kqueue found invalid fd", 0));
|
134
|
+
fd_kill (EV_A_ fd);
|
135
|
+
}
|
136
|
+
}
|
137
|
+
else /* on all other errors, we error out on the fd */
|
138
|
+
{
|
139
|
+
assert (("libev: kqueue found invalid fd", 0));
|
140
|
+
fd_kill (EV_A_ fd);
|
141
|
+
}
|
142
|
+
}
|
143
|
+
}
|
144
|
+
else
|
145
|
+
fd_event (
|
146
|
+
EV_A_
|
147
|
+
fd,
|
148
|
+
kqueue_events [i].filter == EVFILT_READ ? EV_READ
|
149
|
+
: kqueue_events [i].filter == EVFILT_WRITE ? EV_WRITE
|
150
|
+
: 0
|
151
|
+
);
|
152
|
+
}
|
153
|
+
|
154
|
+
if (expect_false (res == kqueue_eventmax))
|
155
|
+
{
|
156
|
+
ev_free (kqueue_events);
|
157
|
+
kqueue_eventmax = array_nextsize (sizeof (struct kevent), kqueue_eventmax, kqueue_eventmax + 1);
|
158
|
+
kqueue_events = (struct kevent *)ev_malloc (sizeof (struct kevent) * kqueue_eventmax);
|
159
|
+
}
|
160
|
+
}
|
161
|
+
|
162
|
+
inline_size
|
163
|
+
int
|
164
|
+
kqueue_init (EV_P_ int flags)
|
165
|
+
{
|
166
|
+
/* initialize the kernel queue */
|
167
|
+
kqueue_fd_pid = getpid ();
|
168
|
+
if ((backend_fd = kqueue ()) < 0)
|
169
|
+
return 0;
|
170
|
+
|
171
|
+
fcntl (backend_fd, F_SETFD, FD_CLOEXEC); /* not sure if necessary, hopefully doesn't hurt */
|
172
|
+
|
173
|
+
backend_mintime = 1e-9; /* apparently, they did the right thing in freebsd */
|
174
|
+
backend_modify = kqueue_modify;
|
175
|
+
backend_poll = kqueue_poll;
|
176
|
+
|
177
|
+
kqueue_eventmax = 64; /* initial number of events receivable per poll */
|
178
|
+
kqueue_events = (struct kevent *)ev_malloc (sizeof (struct kevent) * kqueue_eventmax);
|
179
|
+
|
180
|
+
kqueue_changes = 0;
|
181
|
+
kqueue_changemax = 0;
|
182
|
+
kqueue_changecnt = 0;
|
183
|
+
|
184
|
+
return EVBACKEND_KQUEUE;
|
185
|
+
}
|
186
|
+
|
187
|
+
inline_size
|
188
|
+
void
|
189
|
+
kqueue_destroy (EV_P)
|
190
|
+
{
|
191
|
+
ev_free (kqueue_events);
|
192
|
+
ev_free (kqueue_changes);
|
193
|
+
}
|
194
|
+
|
195
|
+
inline_size
|
196
|
+
void
|
197
|
+
kqueue_fork (EV_P)
|
198
|
+
{
|
199
|
+
/* some BSD kernels don't just destroy the kqueue itself,
|
200
|
+
* but also close the fd, which isn't documented, and
|
201
|
+
* impossible to support properly.
|
202
|
+
* we remember the pid of the kqueue call and only close
|
203
|
+
* the fd if the pid is still the same.
|
204
|
+
* this leaks fds on sane kernels, but BSD interfaces are
|
205
|
+
* notoriously buggy and rarely get fixed.
|
206
|
+
*/
|
207
|
+
pid_t newpid = getpid ();
|
208
|
+
|
209
|
+
if (newpid == kqueue_fd_pid)
|
210
|
+
close (backend_fd);
|
211
|
+
|
212
|
+
kqueue_fd_pid = newpid;
|
213
|
+
while ((backend_fd = kqueue ()) < 0)
|
214
|
+
ev_syserr ("(libev) kqueue");
|
215
|
+
|
216
|
+
fcntl (backend_fd, F_SETFD, FD_CLOEXEC);
|
217
|
+
|
218
|
+
/* re-register interest in fds */
|
219
|
+
fd_rearm_all (EV_A);
|
220
|
+
}
|
221
|
+
|
222
|
+
/* sys/event.h defines EV_ERROR */
|
223
|
+
#undef EV_ERROR
|
224
|
+
|