/* * %CopyrightBegin% * * Copyright Ericsson AB 2010-2014. All Rights Reserved. * * The contents of this file are subject to the Erlang Public License, * Version 1.1, (the "License"); you may not use this file except in * compliance with the License. You should have received a copy of the * Erlang Public License along with this software. If not, it can be * retrieved online at http://www.erlang.org/. * * Software distributed under the License is distributed on an "AS IS" * basis, WITHOUT WARRANTY OF ANY KIND, either express or implied. See * the License for the specific language governing rights and limitations * under the License. * * %CopyrightEnd% */ /* * Description: A Thread library for use in the ERTS and other OTP * applications. * Author: Rickard Green */ #ifdef HAVE_CONFIG_H #include "config.h" #endif #include #define ETHR_INLINE_FUNC_NAME_(X) X ## __ #define ETHR_AUX_IMPL__ #include "ethread.h" #include "ethr_internal.h" #include #include #ifndef __WIN32__ #include #endif #define ERTS_TS_EV_ALLOC_DEFAULT_POOL_SIZE 4000 #define ERTS_TS_EV_ALLOC_POOL_SIZE 25 erts_cpu_info_t *ethr_cpu_info__; int ethr_not_completely_inited__ = 1; int ethr_not_inited__ = 1; ethr_memory_allocators ethr_mem__ = ETHR_MEM_ALLOCS_DEF_INITER__; void *(*ethr_thr_prepare_func__)(void) = NULL; void (*ethr_thr_parent_func__)(void *) = NULL; void (*ethr_thr_child_func__)(void *) = NULL; typedef struct ethr_xhndl_list_ ethr_xhndl_list; struct ethr_xhndl_list_ { ethr_xhndl_list *next; void (*funcp)(void); }; size_t ethr_pagesize__; size_t ethr_min_stack_size__; /* kilo words */ size_t ethr_max_stack_size__; /* kilo words */ ethr_rwmutex xhndl_rwmtx; ethr_xhndl_list *xhndl_list; static int main_threads; static int init_ts_event_alloc(void); ethr_runtime_t ethr_runtime__ #ifdef __GNUC__ __attribute__ ((aligned (ETHR_CACHE_LINE_SIZE))) #endif ; #if defined(ETHR_X86_RUNTIME_CONF__) /* * x86/x86_64 specifics shared between windows and * pthread implementations. */ #define ETHR_IS_X86_VENDOR(V, B, C, D) \ (sizeof(V) == 13 && is_x86_vendor((V), (B), (C), (D))) static ETHR_INLINE int is_x86_vendor(char *str, int ebx, int ecx, int edx) { return (*((int *) &str[0]) == ebx && *((int *) &str[sizeof(int)]) == edx && *((int *) &str[sizeof(int)*2]) == ecx); } static void x86_init(void) { int eax, ebx, ecx, edx; eax = ebx = ecx = edx = 0; ethr_x86_cpuid__(&eax, &ebx, &ecx, &edx); if (eax > 0 && (ETHR_IS_X86_VENDOR("GenuineIntel", ebx, ecx, edx) || ETHR_IS_X86_VENDOR("AuthenticAMD", ebx, ecx, edx))) { eax = 1; ethr_x86_cpuid__(&eax, &ebx, &ecx, &edx); } else { /* * The meaning of the feature flags for this * vendor have not been verified. */ eax = ebx = ecx = edx = 0; } /* * The feature flags tested below have only been verified * for vendors checked above. Also note that only these * feature flags have been verified to have these specific * meanings. If another feature flag test is introduced, * it has to be verified to have the same meaning for all * vendors above. */ #if ETHR_SIZEOF_PTR == 8 /* bit 13 of ecx is set if we have cmpxchg16b */ ethr_runtime__.conf.have_dw_cmpxchg = (ecx & (1 << 13)); #elif ETHR_SIZEOF_PTR == 4 /* bit 8 of edx is set if we have cmpxchg8b */ ethr_runtime__.conf.have_dw_cmpxchg = (edx & (1 << 8)); #else # error "Not supported" #endif /* bit 26 of edx is set if we have sse2 */ ethr_runtime__.conf.have_sse2 = (edx & (1 << 26)); } #endif /* ETHR_X86_RUNTIME_CONF__ */ int ethr_init_common__(ethr_init_data *id) { int res; ethr_init_event__(); #if defined(ETHR_X86_RUNTIME_CONF__) x86_init(); #endif if (id) { ethr_thr_prepare_func__ = id->thread_create_prepare_func; ethr_thr_parent_func__ = id->thread_create_parent_func; ethr_thr_child_func__ = id->thread_create_child_func; } ethr_cpu_info__ = erts_cpu_info_create(); if (!ethr_cpu_info__) return ENOMEM; #ifdef _SC_PAGESIZE ethr_pagesize__ = (size_t) sysconf(_SC_PAGESIZE); #elif defined(HAVE_GETPAGESIZE) ethr_pagesize__ = (size_t) getpagesize(); #else ethr_pagesize__ = (size_t) 4*1024; /* Guess 4 KB */ #endif /* User needs at least 4 KB */ ethr_min_stack_size__ = 4*1024; #if SIZEOF_VOID_P == 8 /* Double that on 64-bit archs */ ethr_min_stack_size__ *= 2; #endif /* On some systems as much as about 4 KB is used by the system */ ethr_min_stack_size__ += 4*1024; /* There should be room for signal handlers */ #ifdef SIGSTKSZ ethr_min_stack_size__ += SIGSTKSZ; #else ethr_min_stack_size__ += ethr_pagesize__; #endif /* The system may think that we need more stack */ #if defined(PTHREAD_STACK_MIN) if (ethr_min_stack_size__ < PTHREAD_STACK_MIN) ethr_min_stack_size__ = PTHREAD_STACK_MIN; #elif defined(_SC_THREAD_STACK_MIN) { size_t thr_min_stk_sz = (size_t) sysconf(_SC_THREAD_STACK_MIN); if (ethr_min_stack_size__ < thr_min_stk_sz) ethr_min_stack_size__ = thr_min_stk_sz; } #endif /* The guard is at least on some platforms included in the stack size passed when creating threads */ #ifdef ETHR_STACK_GUARD_SIZE ethr_min_stack_size__ += ETHR_STACK_GUARD_SIZE; #endif ethr_min_stack_size__ = ETHR_PAGE_ALIGN(ethr_min_stack_size__); ethr_min_stack_size__ = ETHR_B2KW(ethr_min_stack_size__); #ifdef __OSE__ /* For supervisor processes, OSE adds a number of bytes to the requested stack. With this * addition, the resulting size must not exceed the largest available stack size. The number * of bytes that will be added is configured in the monolith and can therefore not be * specified here. We simply assume that it is less than 0x1000. The available stack sizes * are configured in the .lmconf file and the largest one is usually 65536 bytes. * Consequently, the requested stack size is limited to 0xF000. */ ethr_max_stack_size__ = 0xF000; #else ethr_max_stack_size__ = 32*1024*1024; #endif #if SIZEOF_VOID_P == 8 ethr_max_stack_size__ *= 2; #endif ethr_max_stack_size__ = ETHR_B2KW(ethr_max_stack_size__); res = ethr_init_atomics(); if (res != 0) return res; res = ethr_mutex_lib_init(erts_get_cpu_configured(ethr_cpu_info__)); if (res != 0) return res; xhndl_list = NULL; return 0; } int ethr_late_init_common__(ethr_late_init_data *lid) { ethr_ts_event *tsep = NULL; int reader_groups; int res; int i; ethr_memory_allocator *m[] = {ðr_mem__.std, ðr_mem__.sl, ðr_mem__.ll}; if (lid) ethr_mem__ = lid->mem; if (!ethr_mem__.std.alloc || !ethr_mem__.std.realloc || !ethr_mem__.std.free) { ethr_mem__.std.alloc = malloc; ethr_mem__.std.realloc = realloc; ethr_mem__.std.free = free; } for (i = 0; i < sizeof(m)/sizeof(m[0]); i++) { if (!m[i]->alloc || !m[i]->realloc || !m[i]->free) { m[i]->alloc = ethr_mem__.std.alloc; m[i]->realloc = ethr_mem__.std.realloc; m[i]->free = ethr_mem__.std.free; } } res = init_ts_event_alloc(); if (res != 0) return res; res = ethr_make_ts_event__(&tsep); if (res == 0) tsep->iflgs |= ETHR_TS_EV_ETHREAD; if (!lid) { main_threads = 0; reader_groups = 0; } else { if (lid->main_threads < 0 || USHRT_MAX < lid->main_threads) return res; main_threads = lid->main_threads; reader_groups = lid->reader_groups; } res = ethr_mutex_lib_late_init(reader_groups, main_threads); if (res != 0) return res; ethr_not_completely_inited__ = 0; /* Need it for rwmutex_init */ res = ethr_rwmutex_init(&xhndl_rwmtx); ethr_not_completely_inited__ = 1; if (res != 0) return res; return 0; } int ethr_install_exit_handler(void (*funcp)(void)) { ethr_xhndl_list *xhp; #if ETHR_XCHK if (ethr_not_completely_inited__) { ETHR_ASSERT(0); return EACCES; } #endif if (!funcp) return EINVAL; xhp = (ethr_xhndl_list *) ethr_mem__.std.alloc(sizeof(ethr_xhndl_list)); if (!xhp) return ENOMEM; ethr_rwmutex_rwlock(&xhndl_rwmtx); xhp->funcp = funcp; xhp->next = xhndl_list; xhndl_list = xhp; ethr_rwmutex_rwunlock(&xhndl_rwmtx); return 0; } void ethr_run_exit_handlers__(void) { ethr_xhndl_list *xhp; ethr_rwmutex_rlock(&xhndl_rwmtx); xhp = xhndl_list; ethr_rwmutex_runlock(&xhndl_rwmtx); for (; xhp; xhp = xhp->next) (*xhp->funcp)(); } /* * Thread specific event alloc, etc. * * Note that we don't know when it is safe to destroy an event, but * we know when it is safe to reuse it. ts_event_free() therefore * never destroys an event (but makes freed events available for * reuse). * * We could easily keep track of the usage of events, and by this * make it possible to destroy events. We would however suffer a * performance penalty for this and save very little memory. */ typedef union { ethr_ts_event ts_ev; char align[ETHR_CACHE_LINE_ALIGN_SIZE(sizeof(ethr_ts_event))]; } ethr_aligned_ts_event; static ethr_spinlock_t ts_ev_alloc_lock; static ethr_ts_event *free_ts_ev; static ethr_ts_event *ts_event_pool(int size, ethr_ts_event **endpp) { int i; ethr_aligned_ts_event *atsev; atsev = ethr_mem__.std.alloc(sizeof(ethr_aligned_ts_event) * size + ETHR_CACHE_LINE_SIZE - 1); if (!atsev) return NULL; if ((((ethr_uint_t) atsev) & ETHR_CACHE_LINE_MASK) != 0) atsev = ((ethr_aligned_ts_event *) ((((ethr_uint_t) atsev) & ~ETHR_CACHE_LINE_MASK) + ETHR_CACHE_LINE_SIZE)); for (i = 1; i < size; i++) { atsev[i-1].ts_ev.next = &atsev[i].ts_ev; ethr_atomic32_init(&atsev[i-1].ts_ev.uaflgs, 0); atsev[i-1].ts_ev.iflgs = 0; } ethr_atomic32_init(&atsev[size-1].ts_ev.uaflgs, 0); atsev[size-1].ts_ev.iflgs = 0; atsev[size-1].ts_ev.next = NULL; if (endpp) *endpp = &atsev[size-1].ts_ev; return &atsev[0].ts_ev; } static int init_ts_event_alloc(void) { free_ts_ev = ts_event_pool(ERTS_TS_EV_ALLOC_DEFAULT_POOL_SIZE, NULL); if (!free_ts_ev) return ENOMEM; return ethr_spinlock_init(&ts_ev_alloc_lock); } static ethr_ts_event *ts_event_alloc(void) { ethr_ts_event *ts_ev; ethr_spin_lock(&ts_ev_alloc_lock); if (free_ts_ev) { ts_ev = free_ts_ev; free_ts_ev = ts_ev->next; ethr_spin_unlock(&ts_ev_alloc_lock); } else { ethr_ts_event *ts_ev_pool_end; ethr_spin_unlock(&ts_ev_alloc_lock); ts_ev = ts_event_pool(ERTS_TS_EV_ALLOC_POOL_SIZE, &ts_ev_pool_end); if (!ts_ev) return NULL; ethr_spin_lock(&ts_ev_alloc_lock); ts_ev_pool_end->next = free_ts_ev; free_ts_ev = ts_ev->next; ethr_spin_unlock(&ts_ev_alloc_lock); } return ts_ev; } static void ts_event_free(ethr_ts_event *ts_ev) { ETHR_ASSERT(!ts_ev->udata); ethr_spin_lock(&ts_ev_alloc_lock); ts_ev->next = free_ts_ev; free_ts_ev = ts_ev; ethr_spin_unlock(&ts_ev_alloc_lock); } int ethr_make_ts_event__(ethr_ts_event **tsepp) { int res; ethr_ts_event *tsep = *tsepp; if (!tsep) { tsep = ts_event_alloc(); if (!tsep) return ENOMEM; } if ((tsep->iflgs & ETHR_TS_EV_INITED) == 0) { res = ethr_event_init(&tsep->event); if (res != 0) { ts_event_free(tsep); return res; } } tsep->iflgs = ETHR_TS_EV_INITED; tsep->udata = NULL; tsep->rgix = 0; tsep->mtix = 0; res = ethr_set_tse__(tsep); if (res != 0 && tsepp && *tsepp) { ts_event_free(tsep); return res; } if (tsepp) *tsepp = tsep; return 0; } int ethr_get_tmp_ts_event__(ethr_ts_event **tsepp) { int res; ethr_ts_event *tsep = *tsepp; if (!tsep) { tsep = ts_event_alloc(); if (!tsep) return ENOMEM; } if ((tsep->iflgs & ETHR_TS_EV_INITED) == 0) { res = ethr_event_init(&tsep->event); if (res != 0) { ts_event_free(tsep); return res; } } tsep->iflgs = ETHR_TS_EV_INITED|ETHR_TS_EV_TMP; tsep->udata = NULL; if (tsepp) *tsepp = tsep; return 0; } int ethr_free_ts_event__(ethr_ts_event *tsep) { ts_event_free(tsep); return 0; } void ethr_ts_event_destructor__(void *vtsep) { if (vtsep) { ethr_ts_event *tsep = (ethr_ts_event *) vtsep; ts_event_free(tsep); ethr_set_tse__(NULL); } } int ethr_set_main_thr_status(int on, int no) { ethr_ts_event *tsep = ethr_get_tse__(); if (!tsep) return EINVAL; if (on) { if (no < 1 || main_threads < no) return EINVAL; tsep->mtix = (unsigned short) no; tsep->iflgs |= ETHR_TS_EV_MAIN_THR; } else { tsep->iflgs &= ~ETHR_TS_EV_MAIN_THR; tsep->mtix = (unsigned short) 0; } return 0; } int ethr_get_main_thr_status(int *on) { ethr_ts_event *tsep = ethr_get_tse__(); if (!tsep) *on = 0; else { if (tsep->iflgs & ETHR_TS_EV_MAIN_THR) *on = 1; else *on = 0; } return 0; } /* Spinlocks and rwspinlocks */ int ethr_spinlock_init(ethr_spinlock_t *lock) { #if ETHR_XCHK if (!lock) { ETHR_ASSERT(0); return EINVAL; } #endif return ethr_spinlock_init__(lock); } int ethr_spinlock_destroy(ethr_spinlock_t *lock) { #if ETHR_XCHK if (ethr_not_inited__) { ETHR_ASSERT(0); return EACCES; } if (!lock) { ETHR_ASSERT(0); return EINVAL; } #endif return ethr_spinlock_destroy__(lock); } void ethr_spin_unlock(ethr_spinlock_t *lock) { ETHR_ASSERT(!ethr_not_inited__); ETHR_ASSERT(lock); ethr_spin_unlock__(lock); } void ethr_spin_lock(ethr_spinlock_t *lock) { ETHR_ASSERT(!ethr_not_inited__); ETHR_ASSERT(lock); ethr_spin_lock__(lock); } int ethr_rwlock_init(ethr_rwlock_t *lock) { #if ETHR_XCHK if (!lock) { ETHR_ASSERT(0); return EINVAL; } #endif return ethr_rwlock_init__(lock); } int ethr_rwlock_destroy(ethr_rwlock_t *lock) { #if ETHR_XCHK if (ethr_not_inited__) { ETHR_ASSERT(0); return EACCES; } if (!lock) { ETHR_ASSERT(0); return EINVAL; } #endif return ethr_rwlock_destroy__(lock); } void ethr_read_unlock(ethr_rwlock_t *lock) { ETHR_ASSERT(!ethr_not_inited__); ETHR_ASSERT(lock); ethr_read_unlock__(lock); } void ethr_read_lock(ethr_rwlock_t *lock) { ETHR_ASSERT(!ethr_not_inited__); ETHR_ASSERT(lock); ethr_read_lock__(lock); } void ethr_write_unlock(ethr_rwlock_t *lock) { ETHR_ASSERT(!ethr_not_inited__); ETHR_ASSERT(lock); ethr_write_unlock__(lock); } void ethr_write_lock(ethr_rwlock_t *lock) { ETHR_ASSERT(!ethr_not_inited__); ETHR_ASSERT(lock); ethr_write_lock__(lock); } ETHR_IMPL_NORETURN__ ethr_fatal_error__(const char *file, int line, const char *func, int err) { char *errstr; if (err == ENOTSUP) errstr = "Operation not supported"; else { errstr = strerror(err); if (!errstr) errstr = "Unknown error"; } fprintf(stderr, "%s:%d: Fatal error in %s(): %s (%d)\n", file, line, func, errstr, err); ethr_abort__(); } int ethr_assert_failed(const char *file, int line, const char *func, char *a) { fprintf(stderr, "%s:%d: %s(): Assertion failed: %s\n", file, line, func, a); #ifdef __OSE__ ramlog_printf("%d: %s:%d: %s(): Assertion failed: %s\n", current_process(),file, line, func, a); #endif ethr_abort__(); return 0; }