From: Brian Hurt <bhurt@spnz.org>
To: skaller <skaller@users.sourceforge.net>
Cc: William Lovas <wlovas@stwing.upenn.edu>, caml-list@yquem.inria.fr
Subject: Re: [Caml-list] STM support in OCaml
Date: Wed, 8 Mar 2006 16:11:04 -0600 (CST) [thread overview]
Message-ID: <Pine.LNX.4.63.0603081555040.9569@localhost.localdomain> (raw)
In-Reply-To: <1141855594.23909.63.camel@budgie.wigram>
[-- Attachment #1: Type: TEXT/PLAIN, Size: 1608 bytes --]
On Thu, 9 Mar 2006, skaller wrote:
> Ahem. Now try that on an AMDx2 (dual core). The cost goes through
> the roof if one process has a thread on each core. Because each
> core has its own cache and both caches have to be flushed/
> synchronised. And those caches are BIG!
Love to. Wanna buy me the box? :-} Seriously- my code is attached, if
someone wants to run it on other boxes and post the results, feel free.
It's GNU-C/x86 specific, as I'm using GNU C's inline assembler and the
rdtsc instruction to get accurate cycle counts.
As to the cache comment: the whole caches don't have to be flushed, just
the line the mutex is on. Which makes it approximately the cost of a
cache miss- that's a good approximation of the cost of getting an
uncontended lock.
>
> I have no idea if Linux, for example, running SMP kernel,
> is smart enough to know if a mutex is shared between two
> processing units or not: AFAIK Linux doesn't support
> interprocess mutex. Windows does. Be interesting to
> compare.
It doesn't look like the mutex software is even going into the kernel.
I don't think the Linux kernel even knows the mutex *exists*, let alone
what threads are competing for it. On the x86, at least, lock
instructions are not priveledged.
>
> As mentioned before the only data I have at the moment
> is a two thread counter increment experiment on a dual
> CPU G5 box, where the speed up from 2 CPUs vs 1 was
> a factor of 15 .. times SLOWER.
If you're ping-ponging a cache line between two CPUs (and the AMD dual
cores count as two CPUs), then I can easily beleive that.
So?
Brian
[-- Attachment #2: Type: TEXT/X-CSRC, Size: 1847 bytes --]
#include <stdio.h>
#include <pthread.h>
#include <semaphore.h>
#if !defined(__GNUC__) && !defined(__i386__)
#error This code only works with GCC/i386.
#endif
/* The reason this only works under GNU C and the x86 is we're using the
* rdtsc instruction.
*/
static inline unsigned long long rdtsc() {
unsigned long long rval;
asm volatile ("rdtsc" : "=A" (rval));
return rval;
}
static sem_t waiting_thread_semaphore;
static pthread_mutex_t mutex;
void * waiting_thread_func(void * param __attribute__((unused))) {
sem_wait(&waiting_thread_semaphore);
return NULL;
}
int main(void) {
int i;
pthread_t waiting_thread;
void * trash;
unsigned long long start, stop, time, min;
/* Create a thread to force us to actually do multi-threaded work */
sem_init(&waiting_thread_semaphore, 1, 0);
pthread_create(&waiting_thread, NULL, waiting_thread_func, NULL);
pthread_mutex_init(&mutex, NULL);
/* Time how long a rdtsc takes- we do this ten times and take the
* cheapest run.
*/
min = ~0ull;
for (i = 0; i < 10; ++i) {
start = rdtsc();
stop = rdtsc();
time = stop - start;
if (time < min) {
min = time;
}
}
printf("Minimum time for a rdtsc instruction (in clocks): %llu\n", min);
/* Now time how long the pair of mutex lock + unlock take */
min = ~0ull;
for (i = 0; i < 10; ++i) {
start = rdtsc();
pthread_mutex_lock(&mutex);
pthread_mutex_unlock(&mutex);
stop = rdtsc();
time = stop - start;
if (time < min) {
min = time;
}
}
printf("Minimum time for a mutex lock+unlock + rdtsc (in clocks): %llu\n", min);
/* Clean up the waiting thread we spawned just to be multithreaded. */
sem_post(&waiting_thread_semaphore);
pthread_join(waiting_thread, &trash);
sem_destroy(&waiting_thread_semaphore);
return 0;
}
next prev parent reply other threads:[~2006-03-08 22:09 UTC|newest]
Thread overview: 38+ messages / expand[flat|nested] mbox.gz Atom feed top
2006-03-07 16:18 Asfand Yar Qazi
2006-03-07 16:50 ` [Caml-list] " Sebastian Egner
2006-03-07 17:44 ` Michael Hicks
2006-03-08 0:37 ` Asfand Yar Qazi
2006-03-08 5:05 ` Erick Tryzelaar
2006-03-11 19:43 ` Deadlock free locking scheme (was: Re: [Caml-list] STM support in OCaml) David MENTRE
2006-03-07 17:15 ` [Caml-list] STM support in OCaml skaller
2006-03-07 19:05 ` Asfand Yar Qazi
2006-03-08 0:52 ` skaller
2006-03-08 7:08 ` Bardur Arantsson
2006-03-08 10:38 ` [Caml-list] " Asfand Yar Qazi
2006-03-08 19:36 ` William Lovas
2006-03-08 20:45 ` Brian Hurt
2006-03-08 21:14 ` Paul Snively
2006-03-08 22:06 ` skaller
2006-03-08 22:10 ` Gerd Stolpmann
2006-03-08 23:48 ` skaller
2006-03-09 7:45 ` Andrae Muys
2006-03-09 9:18 ` David Brown
2006-03-08 22:11 ` Brian Hurt [this message]
2006-03-08 23:05 ` Lodewijk Vöge
2006-03-09 3:13 ` Brian Hurt
2006-03-08 23:45 ` Robert Roessler
2006-03-09 0:23 ` skaller
2006-03-09 3:19 ` Brian Hurt
2006-03-09 4:32 ` skaller
2006-03-09 10:38 ` John Chu
2006-03-09 16:53 ` Stefan Monnier
2006-03-11 15:26 ` [Caml-list] " Florian Weimer
2006-03-08 10:11 yoann padioleau
2006-03-08 10:41 ` Asfand Yar Qazi
2006-03-08 12:23 ` skaller
2006-03-08 23:02 ` Asfand Yar Qazi
2006-03-09 0:36 ` skaller
2006-03-08 11:32 ` Gerd Stolpmann
2006-03-08 12:04 ` skaller
2006-03-08 19:22 ` Dan Grossman
2006-03-08 22:10 ` skaller
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=Pine.LNX.4.63.0603081555040.9569@localhost.localdomain \
--to=bhurt@spnz.org \
--cc=caml-list@yquem.inria.fr \
--cc=skaller@users.sourceforge.net \
--cc=wlovas@stwing.upenn.edu \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox