diff options
| author | Ulrich Drepper <drepper@redhat.com> | 2000-05-25 06:15:25 +0000 |
|---|---|---|
| committer | Ulrich Drepper <drepper@redhat.com> | 2000-05-25 06:15:25 +0000 |
| commit | 0f5504179a2e37a20e409c48dcc8d640393cd16d (patch) | |
| tree | 5e61a218eacec17393a6688dab8bb0e59943592b /linuxthreads | |
| parent | db33f7d4aef7422140d5e19c440bb5e084fbe186 (diff) | |
| download | glibc-0f5504179a2e37a20e409c48dcc8d640393cd16d.tar.xz glibc-0f5504179a2e37a20e409c48dcc8d640393cd16d.zip | |
Update.
2000-05-23 Jakub Jelinek <jakub@redhat.com>
* sysdeps/i386/fpu/bits/mathinline.h (__sincos, __sincosf,
__sincosl): Guard with __USE_GNU.
Diffstat (limited to 'linuxthreads')
| -rw-r--r-- | linuxthreads/Makefile | 2 | ||||
| -rw-r--r-- | linuxthreads/cancel.c | 6 | ||||
| -rw-r--r-- | linuxthreads/condvar.c | 12 | ||||
| -rw-r--r-- | linuxthreads/internals.h | 4 | ||||
| -rw-r--r-- | linuxthreads/join.c | 24 | ||||
| -rw-r--r-- | linuxthreads/manager.c | 18 | ||||
| -rw-r--r-- | linuxthreads/mutex.c | 6 | ||||
| -rw-r--r-- | linuxthreads/pthread.c | 14 | ||||
| -rw-r--r-- | linuxthreads/rwlock.c | 24 | ||||
| -rw-r--r-- | linuxthreads/semaphore.c | 36 | ||||
| -rw-r--r-- | linuxthreads/signals.c | 4 | ||||
| -rw-r--r-- | linuxthreads/spinlock.c | 34 | ||||
| -rw-r--r-- | linuxthreads/spinlock.h | 8 | ||||
| -rw-r--r-- | linuxthreads/sysdeps/alpha/pspinlock.c | 109 | ||||
| -rw-r--r-- | linuxthreads/sysdeps/arm/pspinlock.c | 81 | ||||
| -rw-r--r-- | linuxthreads/sysdeps/i386/pspinlock.c | 97 | ||||
| -rw-r--r-- | linuxthreads/sysdeps/m68k/pspinlock.c | 81 | ||||
| -rw-r--r-- | linuxthreads/sysdeps/mips/pspinlock.c | 66 | ||||
| -rw-r--r-- | linuxthreads/sysdeps/powerpc/pspinlock.c | 66 | ||||
| -rw-r--r-- | linuxthreads/sysdeps/pthread/bits/pthreadtypes.h | 15 | ||||
| -rw-r--r-- | linuxthreads/sysdeps/sparc/sparc32/pspinlock.c | 66 | ||||
| -rw-r--r-- | linuxthreads/sysdeps/sparc/sparc64/pspinlock.c | 66 |
22 files changed, 723 insertions, 116 deletions
diff --git a/linuxthreads/Makefile b/linuxthreads/Makefile index 6e443631c3..777fb2b408 100644 --- a/linuxthreads/Makefile +++ b/linuxthreads/Makefile @@ -35,7 +35,7 @@ extra-libs-others := $(extra-libs) libpthread-routines := attr cancel condvar join manager mutex ptfork \ ptlongjmp pthread signals specific errno lockfile \ semaphore spinlock wrapsyscall rwlock pt-machine \ - oldsemaphore events getcpuclockid + oldsemaphore events getcpuclockid pspinlock vpath %.c Examples tests = ex1 ex2 ex3 ex4 ex5 ex6 ex7 diff --git a/linuxthreads/cancel.c b/linuxthreads/cancel.c index 067e3f65ed..a51e8ccfc3 100644 --- a/linuxthreads/cancel.c +++ b/linuxthreads/cancel.c @@ -58,14 +58,14 @@ int pthread_cancel(pthread_t thread) __pthread_lock(&handle->h_lock, NULL); if (invalid_handle(handle, thread)) { - __pthread_spin_unlock(&handle->h_lock); + __pthread_unlock(&handle->h_lock); return ESRCH; } th = handle->h_descr; if (th->p_canceled) { - __pthread_spin_unlock(&handle->h_lock); + __pthread_unlock(&handle->h_lock); return 0; } @@ -85,7 +85,7 @@ int pthread_cancel(pthread_t thread) th->p_woken_by_cancel = dorestart; } - __pthread_spin_unlock(&handle->h_lock); + __pthread_unlock(&handle->h_lock); /* If the thread has suspended or is about to, then we unblock it by issuing a restart, instead of a cancel signal. Otherwise we send diff --git a/linuxthreads/condvar.c b/linuxthreads/condvar.c index ab107c1fab..536d88ed05 100644 --- a/linuxthreads/condvar.c +++ b/linuxthreads/condvar.c @@ -50,7 +50,7 @@ static int cond_extricate_func(void *obj, pthread_descr th) __pthread_lock(&cond->__c_lock, self); did_remove = remove_from_queue(&cond->__c_waiting, th); - __pthread_spin_unlock(&cond->__c_lock); + __pthread_unlock(&cond->__c_lock); return did_remove; } @@ -85,7 +85,7 @@ int pthread_cond_wait(pthread_cond_t *cond, pthread_mutex_t *mutex) enqueue(&cond->__c_waiting, self); else already_canceled = 1; - __pthread_spin_unlock(&cond->__c_lock); + __pthread_unlock(&cond->__c_lock); if (already_canceled) { __pthread_set_own_extricate_if(self, 0); @@ -138,7 +138,7 @@ pthread_cond_timedwait_relative(pthread_cond_t *cond, enqueue(&cond->__c_waiting, self); else already_canceled = 1; - __pthread_spin_unlock(&cond->__c_lock); + __pthread_unlock(&cond->__c_lock); if (already_canceled) { __pthread_set_own_extricate_if(self, 0); @@ -155,7 +155,7 @@ pthread_cond_timedwait_relative(pthread_cond_t *cond, __pthread_lock(&cond->__c_lock, self); was_on_queue = remove_from_queue(&cond->__c_waiting, self); - __pthread_spin_unlock(&cond->__c_lock); + __pthread_unlock(&cond->__c_lock); if (was_on_queue) { __pthread_set_own_extricate_if(self, 0); @@ -196,7 +196,7 @@ int pthread_cond_signal(pthread_cond_t *cond) __pthread_lock(&cond->__c_lock, NULL); th = dequeue(&cond->__c_waiting); - __pthread_spin_unlock(&cond->__c_lock); + __pthread_unlock(&cond->__c_lock); if (th != NULL) restart(th); return 0; } @@ -209,7 +209,7 @@ int pthread_cond_broadcast(pthread_cond_t *cond) /* Copy the current state of the waiting queue and empty it */ tosignal = cond->__c_waiting; cond->__c_waiting = NULL; - __pthread_spin_unlock(&cond->__c_lock); + __pthread_unlock(&cond->__c_lock); /* Now signal each process in the queue */ while ((th = dequeue(&tosignal)) != NULL) restart(th); return 0; diff --git a/linuxthreads/internals.h b/linuxthreads/internals.h index e41e5d3a72..3790efed3f 100644 --- a/linuxthreads/internals.h +++ b/linuxthreads/internals.h @@ -138,7 +138,7 @@ struct _pthread_descr_struct { pthread_t p_tid; /* Thread identifier */ int p_pid; /* PID of Unix process */ int p_priority; /* Thread priority (== 0 if not realtime) */ - pthread_spinlock_t * p_lock; /* Spinlock for synchronized accesses */ + struct _pthread_fastlock * p_lock; /* Spinlock for synchronized accesses */ int p_signal; /* last signal received */ sigjmp_buf * p_signal_jmp; /* where to siglongjmp on a signal or NULL */ sigjmp_buf * p_cancel_jmp; /* where to siglongjmp on a cancel or NULL */ @@ -189,7 +189,7 @@ struct _pthread_descr_struct { typedef struct pthread_handle_struct * pthread_handle; struct pthread_handle_struct { - pthread_spinlock_t h_lock; /* Fast lock for sychronized access */ + struct _pthread_fastlock h_lock; /* Fast lock for sychronized access */ pthread_descr h_descr; /* Thread descriptor or NULL if invalid */ char * h_bottom; /* Lowest address in the stack thread */ }; diff --git a/linuxthreads/join.c b/linuxthreads/join.c index 7c9b6c5fd3..95c0ab6963 100644 --- a/linuxthreads/join.c +++ b/linuxthreads/join.c @@ -62,7 +62,7 @@ void pthread_exit(void * retval) THREAD_SETMEM(self, p_terminated, 1); /* See if someone is joining on us */ joining = THREAD_GETMEM(self, p_joining); - __pthread_spin_unlock(THREAD_GETMEM(self, p_lock)); + __pthread_unlock(THREAD_GETMEM(self, p_lock)); /* Restart joining thread if any */ if (joining != NULL) restart(joining); /* If this is the initial thread, block until all threads have terminated. @@ -76,7 +76,7 @@ void pthread_exit(void * retval) /* Main thread flushes stdio streams and runs atexit functions. It also calls a handler within LinuxThreads which sends a process exit request to the thread manager. */ - exit(0); + exit(0); } /* Threads other than the main one terminate without flushing stdio streams or running atexit functions. */ @@ -97,7 +97,7 @@ static int join_extricate_func(void *obj, pthread_descr th) jo = handle->h_descr; did_remove = jo->p_joining != NULL; jo->p_joining = NULL; - __pthread_spin_unlock(&handle->h_lock); + __pthread_unlock(&handle->h_lock); return did_remove; } @@ -117,17 +117,17 @@ int pthread_join(pthread_t thread_id, void ** thread_return) __pthread_lock(&handle->h_lock, self); if (invalid_handle(handle, thread_id)) { - __pthread_spin_unlock(&handle->h_lock); + __pthread_unlock(&handle->h_lock); return ESRCH; } th = handle->h_descr; if (th == self) { - __pthread_spin_unlock(&handle->h_lock); + __pthread_unlock(&handle->h_lock); return EDEADLK; } /* If detached or already joined, error */ if (th->p_detached || th->p_joining != NULL) { - __pthread_spin_unlock(&handle->h_lock); + __pthread_unlock(&handle->h_lock); return EINVAL; } /* If not terminated yet, suspend ourselves. */ @@ -139,7 +139,7 @@ int pthread_join(pthread_t thread_id, void ** thread_return) th->p_joining = self; else already_canceled = 1; - __pthread_spin_unlock(&handle->h_lock); + __pthread_unlock(&handle->h_lock); if (already_canceled) { __pthread_set_own_extricate_if(self, 0); @@ -160,7 +160,7 @@ int pthread_join(pthread_t thread_id, void ** thread_return) } /* Get return value */ if (thread_return != NULL) *thread_return = th->p_retval; - __pthread_spin_unlock(&handle->h_lock); + __pthread_unlock(&handle->h_lock); /* Send notification to thread manager */ if (__pthread_manager_request >= 0) { request.req_thread = self; @@ -181,24 +181,24 @@ int pthread_detach(pthread_t thread_id) __pthread_lock(&handle->h_lock, NULL); if (invalid_handle(handle, thread_id)) { - __pthread_spin_unlock(&handle->h_lock); + __pthread_unlock(&handle->h_lock); return ESRCH; } th = handle->h_descr; /* If already detached, error */ if (th->p_detached) { - __pthread_spin_unlock(&handle->h_lock); + __pthread_unlock(&handle->h_lock); return EINVAL; } /* If already joining, don't do anything. */ if (th->p_joining != NULL) { - __pthread_spin_unlock(&handle->h_lock); + __pthread_unlock(&handle->h_lock); return 0; } /* Mark as detached */ th->p_detached = 1; terminated = th->p_terminated; - __pthread_spin_unlock(&handle->h_lock); + __pthread_unlock(&handle->h_lock); /* If already terminated, notify thread manager to reclaim resources */ if (terminated && __pthread_manager_request >= 0) { request.req_thread = thread_self(); diff --git a/linuxthreads/manager.c b/linuxthreads/manager.c index 2cc652df97..149cc938b9 100644 --- a/linuxthreads/manager.c +++ b/linuxthreads/manager.c @@ -177,7 +177,7 @@ int __pthread_manager(void *arg) __on_exit handler, which in turn will send REQ_PROCESS_EXIT to the thread manager. In case you are wondering how the manager terminates from its loop here. */ - } + } break; case REQ_POST: __new_sem_post(request.req_args.post); @@ -207,7 +207,7 @@ int __pthread_manager_event(void *arg) /* Get the lock the manager will free once all is correctly set up. */ __pthread_lock (THREAD_GETMEM((&__pthread_manager_thread), p_lock), NULL); /* Free it immediately. */ - __pthread_spin_unlock (THREAD_GETMEM((&__pthread_manager_thread), p_lock)); + __pthread_unlock (THREAD_GETMEM((&__pthread_manager_thread), p_lock)); return __pthread_manager(arg); } @@ -273,7 +273,7 @@ static int pthread_start_thread_event(void *arg) /* Get the lock the manager will free once all is correctly set up. */ __pthread_lock (THREAD_GETMEM(self, p_lock), NULL); /* Free it immediately. */ - __pthread_spin_unlock (THREAD_GETMEM(self, p_lock)); + __pthread_unlock (THREAD_GETMEM(self, p_lock)); /* Continue with the real function. */ return pthread_start_thread (arg); @@ -474,7 +474,7 @@ static int pthread_handle_create(pthread_t *thread, const pthread_attr_t *attr, __linuxthreads_create_event (); /* Now restart the thread. */ - __pthread_spin_unlock(new_thread->p_lock); + __pthread_unlock(new_thread->p_lock); } } } @@ -523,7 +523,7 @@ static void pthread_free(pthread_descr th) __pthread_lock(&handle->h_lock, NULL); handle->h_descr = NULL; handle->h_bottom = (char *)(-1L); - __pthread_spin_unlock(&handle->h_lock); + __pthread_unlock(&handle->h_lock); #ifdef FREE_THREAD FREE_THREAD(th, th->p_nr); #endif @@ -594,7 +594,7 @@ static void pthread_exited(pid_t pid) } } detached = th->p_detached; - __pthread_spin_unlock(th->p_lock); + __pthread_unlock(th->p_lock); if (detached) pthread_free(th); break; @@ -637,19 +637,19 @@ static void pthread_handle_free(pthread_t th_id) if (invalid_handle(handle, th_id)) { /* pthread_reap_children has deallocated the thread already, nothing needs to be done */ - __pthread_spin_unlock(&handle->h_lock); + __pthread_unlock(&handle->h_lock); return; } th = handle->h_descr; if (th->p_exited) { - __pthread_spin_unlock(&handle->h_lock); + __pthread_unlock(&handle->h_lock); pthread_free(th); } else { /* The Unix process of the thread is still running. Mark the thread as detached so that the thread manager will deallocate its resources when the Unix process exits. */ th->p_detached = 1; - __pthread_spin_unlock(&handle->h_lock); + __pthread_unlock(&handle->h_lock); } } diff --git a/linuxthreads/mutex.c b/linuxthreads/mutex.c index a42167cce0..6494323006 100644 --- a/linuxthreads/mutex.c +++ b/linuxthreads/mutex.c @@ -110,7 +110,7 @@ int __pthread_mutex_unlock(pthread_mutex_t * mutex) { switch (mutex->__m_kind) { case PTHREAD_MUTEX_FAST_NP: - __pthread_spin_unlock(&mutex->__m_lock); + __pthread_unlock(&mutex->__m_lock); return 0; case PTHREAD_MUTEX_RECURSIVE_NP: if (mutex->__m_count > 0) { @@ -118,13 +118,13 @@ int __pthread_mutex_unlock(pthread_mutex_t * mutex) return 0; } mutex->__m_owner = NULL; - __pthread_spin_unlock(&mutex->__m_lock); + __pthread_unlock(&mutex->__m_lock); return 0; case PTHREAD_MUTEX_ERRORCHECK_NP: if (mutex->__m_owner != thread_self() || mutex->__m_lock.__status == 0) return EPERM; mutex->__m_owner = NULL; - __pthread_spin_unlock(&mutex->__m_lock); + __pthread_unlock(&mutex->__m_lock); return 0; default: return EINVAL; diff --git a/linuxthreads/pthread.c b/linuxthreads/pthread.c index 0e713e74da..0f63127475 100644 --- a/linuxthreads/pthread.c +++ b/linuxthreads/pthread.c @@ -460,7 +460,7 @@ int __pthread_initialize_manager(void) __linuxthreads_create_event (); /* Now restart the thread. */ - __pthread_spin_unlock(__pthread_manager_thread.p_lock); + __pthread_unlock(__pthread_manager_thread.p_lock); } } } @@ -587,16 +587,16 @@ int pthread_setschedparam(pthread_t thread, int policy, __pthread_lock(&handle->h_lock, NULL); if (invalid_handle(handle, thread)) { - __pthread_spin_unlock(&handle->h_lock); + __pthread_unlock(&handle->h_lock); return ESRCH; } th = handle->h_descr; if (__sched_setscheduler(th->p_pid, policy, param) == -1) { - __pthread_spin_unlock(&handle->h_lock); + __pthread_unlock(&handle->h_lock); return errno; } th->p_priority = policy == SCHED_OTHER ? 0 : param->sched_priority; - __pthread_spin_unlock(&handle->h_lock); + __pthread_unlock(&handle->h_lock); if (__pthread_manager_request >= 0) __pthread_manager_adjust_prio(th->p_priority); return 0; @@ -610,11 +610,11 @@ int pthread_getschedparam(pthread_t thread, int *policy, __pthread_lock(&handle->h_lock, NULL); if (invalid_handle(handle, thread)) { - __pthread_spin_unlock(&handle->h_lock); + __pthread_unlock(&handle->h_lock); return ESRCH; } pid = handle->h_descr->p_pid; - __pthread_spin_unlock(&handle->h_lock); + __pthread_unlock(&handle->h_lock); pol = __sched_getscheduler(pid); if (pol == -1) return errno; if (__sched_getparam(pid, param) == -1) return errno; @@ -788,7 +788,7 @@ void __pthread_set_own_extricate_if(pthread_descr self, pthread_extricate_if *pe { __pthread_lock(self->p_lock, self); THREAD_SETMEM(self, p_extricate, peif); - __pthread_spin_unlock(self->p_lock); + __pthread_unlock(self->p_lock); } /* Primitives for controlling thread execution */ diff --git a/linuxthreads/rwlock.c b/linuxthreads/rwlock.c index e4a4c81f8c..9da87d25d1 100644 --- a/linuxthreads/rwlock.c +++ b/linuxthreads/rwlock.c @@ -217,7 +217,7 @@ __pthread_rwlock_destroy (pthread_rwlock_t *rwlock) __pthread_lock (&rwlock->__rw_lock, NULL); readers = rwlock->__rw_readers; writer = rwlock->__rw_writer; - __pthread_spin_unlock (&rwlock->__rw_lock); + __pthread_unlock (&rwlock->__rw_lock); if (readers > 0 || writer != NULL) return EBUSY; @@ -247,12 +247,12 @@ __pthread_rwlock_rdlock (pthread_rwlock_t *rwlock) break; enqueue (&rwlock->__rw_read_waiting, self); - __pthread_spin_unlock (&rwlock->__rw_lock); + __pthread_unlock (&rwlock->__rw_lock); suspend (self); /* This is not a cancellation point */ } ++rwlock->__rw_readers; - __pthread_spin_unlock (&rwlock->__rw_lock); + __pthread_unlock (&rwlock->__rw_lock); if (have_lock_already || out_of_mem) { @@ -291,7 +291,7 @@ __pthread_rwlock_tryrdlock (pthread_rwlock_t *rwlock) retval = 0; } - __pthread_spin_unlock (&rwlock->__rw_lock); + __pthread_unlock (&rwlock->__rw_lock); if (retval == 0) { @@ -320,13 +320,13 @@ __pthread_rwlock_wrlock (pthread_rwlock_t *rwlock) if (rwlock->__rw_readers == 0 && rwlock->__rw_writer == NULL) { rwlock->__rw_writer = self; - __pthread_spin_unlock (&rwlock->__rw_lock); + __pthread_unlock (&rwlock->__rw_lock); return 0; } /* Suspend ourselves, then try again */ enqueue (&rwlock->__rw_write_waiting, self); - __pthread_spin_unlock (&rwlock->__rw_lock); + __pthread_unlock (&rwlock->__rw_lock); suspend (self); /* This is not a cancellation point */ } } @@ -344,7 +344,7 @@ __pthread_rwlock_trywrlock (pthread_rwlock_t *rwlock) rwlock->__rw_writer = thread_self (); result = 0; } - __pthread_spin_unlock (&rwlock->__rw_lock); + __pthread_unlock (&rwlock->__rw_lock); return result; } @@ -363,7 +363,7 @@ __pthread_rwlock_unlock (pthread_rwlock_t *rwlock) /* Unlocking a write lock. */ if (rwlock->__rw_writer != thread_self ()) { - __pthread_spin_unlock (&rwlock->__rw_lock); + __pthread_unlock (&rwlock->__rw_lock); return EPERM; } rwlock->__rw_writer = NULL; @@ -375,14 +375,14 @@ __pthread_rwlock_unlock (pthread_rwlock_t *rwlock) /* Restart all waiting readers. */ torestart = rwlock->__rw_read_waiting; rwlock->__rw_read_waiting = NULL; - __pthread_spin_unlock (&rwlock->__rw_lock); + __pthread_unlock (&rwlock->__rw_lock); while ((th = dequeue (&torestart)) != NULL) restart (th); } else { /* Restart one waiting writer. */ - __pthread_spin_unlock (&rwlock->__rw_lock); + __pthread_unlock (&rwlock->__rw_lock); restart (th); } } @@ -391,7 +391,7 @@ __pthread_rwlock_unlock (pthread_rwlock_t *rwlock) /* Unlocking a read lock. */ if (rwlock->__rw_readers == 0) { - __pthread_spin_unlock (&rwlock-&g |
