]> CyberLeo.Net >> Repos - FreeBSD/FreeBSD.git/blob - contrib/libevent/evthread.c
amd64: use register macros for gdb_cpu_getreg()
[FreeBSD/FreeBSD.git] / contrib / libevent / evthread.c
1 /*
2  * Copyright (c) 2008-2012 Niels Provos, Nick Mathewson
3  *
4  * Redistribution and use in source and binary forms, with or without
5  * modification, are permitted provided that the following conditions
6  * are met:
7  * 1. Redistributions of source code must retain the above copyright
8  *    notice, this list of conditions and the following disclaimer.
9  * 2. Redistributions in binary form must reproduce the above copyright
10  *    notice, this list of conditions and the following disclaimer in the
11  *    documentation and/or other materials provided with the distribution.
12  * 3. The name of the author may not be used to endorse or promote products
13  *    derived from this software without specific prior written permission.
14  *
15  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
16  * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
17  * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
18  * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
19  * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
20  * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
21  * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
22  * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
23  * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
24  * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
25  */
26
27 #include "event2/event-config.h"
28 #include "evconfig-private.h"
29
30 #ifndef EVENT__DISABLE_THREAD_SUPPORT
31
32 #include "event2/thread.h"
33
34 #include <stdlib.h>
35 #include <string.h>
36
37 #include "log-internal.h"
38 #include "mm-internal.h"
39 #include "util-internal.h"
40 #include "evthread-internal.h"
41
42 #ifdef EVTHREAD_EXPOSE_STRUCTS
43 #define GLOBAL
44 #else
45 #define GLOBAL static
46 #endif
47
48 #ifndef EVENT__DISABLE_DEBUG_MODE
49 extern int event_debug_created_threadable_ctx_;                    
50 extern int event_debug_mode_on_;
51 #endif
52
53 /* globals */
54 GLOBAL int evthread_lock_debugging_enabled_ = 0;
55 GLOBAL struct evthread_lock_callbacks evthread_lock_fns_ = {
56         0, 0, NULL, NULL, NULL, NULL
57 };
58 GLOBAL unsigned long (*evthread_id_fn_)(void) = NULL;
59 GLOBAL struct evthread_condition_callbacks evthread_cond_fns_ = {
60         0, NULL, NULL, NULL, NULL
61 };
62
63 /* Used for debugging */
64 static struct evthread_lock_callbacks original_lock_fns_ = {
65         0, 0, NULL, NULL, NULL, NULL
66 };
67 static struct evthread_condition_callbacks original_cond_fns_ = {
68         0, NULL, NULL, NULL, NULL
69 };
70
71 void
72 evthread_set_id_callback(unsigned long (*id_fn)(void))
73 {
74         evthread_id_fn_ = id_fn;
75 }
76
77 struct evthread_lock_callbacks *evthread_get_lock_callbacks()
78 {
79         return evthread_lock_debugging_enabled_
80             ? &original_lock_fns_ : &evthread_lock_fns_;
81 }
82 struct evthread_condition_callbacks *evthread_get_condition_callbacks()
83 {
84         return evthread_lock_debugging_enabled_
85             ? &original_cond_fns_ : &evthread_cond_fns_;
86 }
87 void evthreadimpl_disable_lock_debugging_(void)
88 {
89         evthread_lock_debugging_enabled_ = 0;
90 }
91
92 int
93 evthread_set_lock_callbacks(const struct evthread_lock_callbacks *cbs)
94 {
95         struct evthread_lock_callbacks *target = evthread_get_lock_callbacks();
96
97 #ifndef EVENT__DISABLE_DEBUG_MODE
98         if (event_debug_mode_on_) {
99                 if (event_debug_created_threadable_ctx_) {
100                     event_errx(1, "evthread initialization must be called BEFORE anything else!");
101                 }
102         }
103 #endif
104
105         if (!cbs) {
106                 if (target->alloc)
107                         event_warnx("Trying to disable lock functions after "
108                             "they have been set up will probaby not work.");
109                 memset(target, 0, sizeof(evthread_lock_fns_));
110                 return 0;
111         }
112         if (target->alloc) {
113                 /* Uh oh; we already had locking callbacks set up.*/
114                 if (target->lock_api_version == cbs->lock_api_version &&
115                         target->supported_locktypes == cbs->supported_locktypes &&
116                         target->alloc == cbs->alloc &&
117                         target->free == cbs->free &&
118                         target->lock == cbs->lock &&
119                         target->unlock == cbs->unlock) {
120                         /* no change -- allow this. */
121                         return 0;
122                 }
123                 event_warnx("Can't change lock callbacks once they have been "
124                     "initialized.");
125                 return -1;
126         }
127         if (cbs->alloc && cbs->free && cbs->lock && cbs->unlock) {
128                 memcpy(target, cbs, sizeof(evthread_lock_fns_));
129                 return event_global_setup_locks_(1);
130         } else {
131                 return -1;
132         }
133 }
134
135 int
136 evthread_set_condition_callbacks(const struct evthread_condition_callbacks *cbs)
137 {
138         struct evthread_condition_callbacks *target = evthread_get_condition_callbacks();
139
140 #ifndef EVENT__DISABLE_DEBUG_MODE
141         if (event_debug_mode_on_) {
142                 if (event_debug_created_threadable_ctx_) {
143                     event_errx(1, "evthread initialization must be called BEFORE anything else!");
144                 }
145         }
146 #endif
147
148         if (!cbs) {
149                 if (target->alloc_condition)
150                         event_warnx("Trying to disable condition functions "
151                             "after they have been set up will probaby not "
152                             "work.");
153                 memset(target, 0, sizeof(evthread_cond_fns_));
154                 return 0;
155         }
156         if (target->alloc_condition) {
157                 /* Uh oh; we already had condition callbacks set up.*/
158                 if (target->condition_api_version == cbs->condition_api_version &&
159                         target->alloc_condition == cbs->alloc_condition &&
160                         target->free_condition == cbs->free_condition &&
161                         target->signal_condition == cbs->signal_condition &&
162                         target->wait_condition == cbs->wait_condition) {
163                         /* no change -- allow this. */
164                         return 0;
165                 }
166                 event_warnx("Can't change condition callbacks once they "
167                     "have been initialized.");
168                 return -1;
169         }
170         if (cbs->alloc_condition && cbs->free_condition &&
171             cbs->signal_condition && cbs->wait_condition) {
172                 memcpy(target, cbs, sizeof(evthread_cond_fns_));
173         }
174         if (evthread_lock_debugging_enabled_) {
175                 evthread_cond_fns_.alloc_condition = cbs->alloc_condition;
176                 evthread_cond_fns_.free_condition = cbs->free_condition;
177                 evthread_cond_fns_.signal_condition = cbs->signal_condition;
178         }
179         return 0;
180 }
181
182 #define DEBUG_LOCK_SIG  0xdeb0b10c
183
184 struct debug_lock {
185         unsigned signature;
186         unsigned locktype;
187         unsigned long held_by;
188         /* XXXX if we ever use read-write locks, we will need a separate
189          * lock to protect count. */
190         int count;
191         void *lock;
192 };
193
194 static void *
195 debug_lock_alloc(unsigned locktype)
196 {
197         struct debug_lock *result = mm_malloc(sizeof(struct debug_lock));
198         if (!result)
199                 return NULL;
200         if (original_lock_fns_.alloc) {
201                 if (!(result->lock = original_lock_fns_.alloc(
202                                 locktype|EVTHREAD_LOCKTYPE_RECURSIVE))) {
203                         mm_free(result);
204                         return NULL;
205                 }
206         } else {
207                 result->lock = NULL;
208         }
209         result->signature = DEBUG_LOCK_SIG;
210         result->locktype = locktype;
211         result->count = 0;
212         result->held_by = 0;
213         return result;
214 }
215
216 static void
217 debug_lock_free(void *lock_, unsigned locktype)
218 {
219         struct debug_lock *lock = lock_;
220         EVUTIL_ASSERT(lock->count == 0);
221         EVUTIL_ASSERT(locktype == lock->locktype);
222         EVUTIL_ASSERT(DEBUG_LOCK_SIG == lock->signature);
223         if (original_lock_fns_.free) {
224                 original_lock_fns_.free(lock->lock,
225                     lock->locktype|EVTHREAD_LOCKTYPE_RECURSIVE);
226         }
227         lock->lock = NULL;
228         lock->count = -100;
229         lock->signature = 0x12300fda;
230         mm_free(lock);
231 }
232
233 static void
234 evthread_debug_lock_mark_locked(unsigned mode, struct debug_lock *lock)
235 {
236         EVUTIL_ASSERT(DEBUG_LOCK_SIG == lock->signature);
237         ++lock->count;
238         if (!(lock->locktype & EVTHREAD_LOCKTYPE_RECURSIVE))
239                 EVUTIL_ASSERT(lock->count == 1);
240         if (evthread_id_fn_) {
241                 unsigned long me;
242                 me = evthread_id_fn_();
243                 if (lock->count > 1)
244                         EVUTIL_ASSERT(lock->held_by == me);
245                 lock->held_by = me;
246         }
247 }
248
249 static int
250 debug_lock_lock(unsigned mode, void *lock_)
251 {
252         struct debug_lock *lock = lock_;
253         int res = 0;
254         if (lock->locktype & EVTHREAD_LOCKTYPE_READWRITE)
255                 EVUTIL_ASSERT(mode & (EVTHREAD_READ|EVTHREAD_WRITE));
256         else
257                 EVUTIL_ASSERT((mode & (EVTHREAD_READ|EVTHREAD_WRITE)) == 0);
258         if (original_lock_fns_.lock)
259                 res = original_lock_fns_.lock(mode, lock->lock);
260         if (!res) {
261                 evthread_debug_lock_mark_locked(mode, lock);
262         }
263         return res;
264 }
265
266 static void
267 evthread_debug_lock_mark_unlocked(unsigned mode, struct debug_lock *lock)
268 {
269         EVUTIL_ASSERT(DEBUG_LOCK_SIG == lock->signature);
270         if (lock->locktype & EVTHREAD_LOCKTYPE_READWRITE)
271                 EVUTIL_ASSERT(mode & (EVTHREAD_READ|EVTHREAD_WRITE));
272         else
273                 EVUTIL_ASSERT((mode & (EVTHREAD_READ|EVTHREAD_WRITE)) == 0);
274         if (evthread_id_fn_) {
275                 unsigned long me;
276                 me = evthread_id_fn_();
277                 EVUTIL_ASSERT(lock->held_by == me);
278                 if (lock->count == 1)
279                         lock->held_by = 0;
280         }
281         --lock->count;
282         EVUTIL_ASSERT(lock->count >= 0);
283 }
284
285 static int
286 debug_lock_unlock(unsigned mode, void *lock_)
287 {
288         struct debug_lock *lock = lock_;
289         int res = 0;
290         evthread_debug_lock_mark_unlocked(mode, lock);
291         if (original_lock_fns_.unlock)
292                 res = original_lock_fns_.unlock(mode, lock->lock);
293         return res;
294 }
295
296 static int
297 debug_cond_wait(void *cond_, void *lock_, const struct timeval *tv)
298 {
299         int r;
300         struct debug_lock *lock = lock_;
301         EVUTIL_ASSERT(lock);
302         EVUTIL_ASSERT(DEBUG_LOCK_SIG == lock->signature);
303         EVLOCK_ASSERT_LOCKED(lock_);
304         evthread_debug_lock_mark_unlocked(0, lock);
305         r = original_cond_fns_.wait_condition(cond_, lock->lock, tv);
306         evthread_debug_lock_mark_locked(0, lock);
307         return r;
308 }
309
310 /* misspelled version for backward compatibility */
311 void
312 evthread_enable_lock_debuging(void)
313 {
314         evthread_enable_lock_debugging();
315 }
316
317 void
318 evthread_enable_lock_debugging(void)
319 {
320         struct evthread_lock_callbacks cbs = {
321                 EVTHREAD_LOCK_API_VERSION,
322                 EVTHREAD_LOCKTYPE_RECURSIVE,
323                 debug_lock_alloc,
324                 debug_lock_free,
325                 debug_lock_lock,
326                 debug_lock_unlock
327         };
328         if (evthread_lock_debugging_enabled_)
329                 return;
330         memcpy(&original_lock_fns_, &evthread_lock_fns_,
331             sizeof(struct evthread_lock_callbacks));
332         memcpy(&evthread_lock_fns_, &cbs,
333             sizeof(struct evthread_lock_callbacks));
334
335         memcpy(&original_cond_fns_, &evthread_cond_fns_,
336             sizeof(struct evthread_condition_callbacks));
337         evthread_cond_fns_.wait_condition = debug_cond_wait;
338         evthread_lock_debugging_enabled_ = 1;
339
340         /* XXX return value should get checked. */
341         event_global_setup_locks_(0);
342 }
343
344 int
345 evthread_is_debug_lock_held_(void *lock_)
346 {
347         struct debug_lock *lock = lock_;
348         if (! lock->count)
349                 return 0;
350         if (evthread_id_fn_) {
351                 unsigned long me = evthread_id_fn_();
352                 if (lock->held_by != me)
353                         return 0;
354         }
355         return 1;
356 }
357
358 void *
359 evthread_debug_get_real_lock_(void *lock_)
360 {
361         struct debug_lock *lock = lock_;
362         return lock->lock;
363 }
364
365 void *
366 evthread_setup_global_lock_(void *lock_, unsigned locktype, int enable_locks)
367 {
368         /* there are four cases here:
369            1) we're turning on debugging; locking is not on.
370            2) we're turning on debugging; locking is on.
371            3) we're turning on locking; debugging is not on.
372            4) we're turning on locking; debugging is on. */
373
374         if (!enable_locks && original_lock_fns_.alloc == NULL) {
375                 /* Case 1: allocate a debug lock. */
376                 EVUTIL_ASSERT(lock_ == NULL);
377                 return debug_lock_alloc(locktype);
378         } else if (!enable_locks && original_lock_fns_.alloc != NULL) {
379                 /* Case 2: wrap the lock in a debug lock. */
380                 struct debug_lock *lock;
381                 EVUTIL_ASSERT(lock_ != NULL);
382
383                 if (!(locktype & EVTHREAD_LOCKTYPE_RECURSIVE)) {
384                         /* We can't wrap it: We need a recursive lock */
385                         original_lock_fns_.free(lock_, locktype);
386                         return debug_lock_alloc(locktype);
387                 }
388                 lock = mm_malloc(sizeof(struct debug_lock));
389                 if (!lock) {
390                         original_lock_fns_.free(lock_, locktype);
391                         return NULL;
392                 }
393                 lock->lock = lock_;
394                 lock->locktype = locktype;
395                 lock->count = 0;
396                 lock->held_by = 0;
397                 return lock;
398         } else if (enable_locks && ! evthread_lock_debugging_enabled_) {
399                 /* Case 3: allocate a regular lock */
400                 EVUTIL_ASSERT(lock_ == NULL);
401                 return evthread_lock_fns_.alloc(locktype);
402         } else {
403                 /* Case 4: Fill in a debug lock with a real lock */
404                 struct debug_lock *lock = lock_ ? lock_ : debug_lock_alloc(locktype);
405                 EVUTIL_ASSERT(enable_locks &&
406                               evthread_lock_debugging_enabled_);
407                 EVUTIL_ASSERT(lock->locktype == locktype);
408                 if (!lock->lock) {
409                         lock->lock = original_lock_fns_.alloc(
410                                 locktype|EVTHREAD_LOCKTYPE_RECURSIVE);
411                         if (!lock->lock) {
412                                 lock->count = -200;
413                                 mm_free(lock);
414                                 return NULL;
415                         }
416                 }
417                 return lock;
418         }
419 }
420
421
422 #ifndef EVTHREAD_EXPOSE_STRUCTS
423 unsigned long
424 evthreadimpl_get_id_()
425 {
426         return evthread_id_fn_ ? evthread_id_fn_() : 1;
427 }
428 void *
429 evthreadimpl_lock_alloc_(unsigned locktype)
430 {
431 #ifndef EVENT__DISABLE_DEBUG_MODE
432         if (event_debug_mode_on_) {
433                 event_debug_created_threadable_ctx_ = 1;
434         }
435 #endif
436
437         return evthread_lock_fns_.alloc ?
438             evthread_lock_fns_.alloc(locktype) : NULL;
439 }
440 void
441 evthreadimpl_lock_free_(void *lock, unsigned locktype)
442 {
443         if (evthread_lock_fns_.free)
444                 evthread_lock_fns_.free(lock, locktype);
445 }
446 int
447 evthreadimpl_lock_lock_(unsigned mode, void *lock)
448 {
449         if (evthread_lock_fns_.lock)
450                 return evthread_lock_fns_.lock(mode, lock);
451         else
452                 return 0;
453 }
454 int
455 evthreadimpl_lock_unlock_(unsigned mode, void *lock)
456 {
457         if (evthread_lock_fns_.unlock)
458                 return evthread_lock_fns_.unlock(mode, lock);
459         else
460                 return 0;
461 }
462 void *
463 evthreadimpl_cond_alloc_(unsigned condtype)
464 {
465 #ifndef EVENT__DISABLE_DEBUG_MODE
466         if (event_debug_mode_on_) {
467                 event_debug_created_threadable_ctx_ = 1;
468         }
469 #endif
470
471         return evthread_cond_fns_.alloc_condition ?
472             evthread_cond_fns_.alloc_condition(condtype) : NULL;
473 }
474 void
475 evthreadimpl_cond_free_(void *cond)
476 {
477         if (evthread_cond_fns_.free_condition)
478                 evthread_cond_fns_.free_condition(cond);
479 }
480 int
481 evthreadimpl_cond_signal_(void *cond, int broadcast)
482 {
483         if (evthread_cond_fns_.signal_condition)
484                 return evthread_cond_fns_.signal_condition(cond, broadcast);
485         else
486                 return 0;
487 }
488 int
489 evthreadimpl_cond_wait_(void *cond, void *lock, const struct timeval *tv)
490 {
491         if (evthread_cond_fns_.wait_condition)
492                 return evthread_cond_fns_.wait_condition(cond, lock, tv);
493         else
494                 return 0;
495 }
496 int
497 evthreadimpl_is_lock_debugging_enabled_(void)
498 {
499         return evthread_lock_debugging_enabled_;
500 }
501
502 int
503 evthreadimpl_locking_enabled_(void)
504 {
505         return evthread_lock_fns_.lock != NULL;
506 }
507 #endif
508
509 #endif