<< Chapter < Page Chapter >> Page >

In all of the above examples, we have focused on the mechanics of shared memory, thread creation, and thread termination. We have used the sleep( ) routine to slow things down sufficiently to see interactions between processes. But we want to go very fast, not just learn threading for threading’s sake.

The example code below uses the multithreading techniques described in this chapter to speed up a sum of a large array. The hpcwall routine is from [link] .

This code allocates a four-million-element double-precision array and fills it with random numbers between 0 and 1. Then using one, two, three, and four threads, it sums up the elements in the array:


#define _REENTRANT /* basic 3-lines for threads */ #include<stdio.h>#include<stdlib.h>#include<pthread.h>#define MAX_THREAD 4 void *SumFunc(void *);int ThreadCount; /* Threads on this try */ double GlobSum; /* A global variable */int index[MAX_THREAD]; /* Local zero-based thread index */pthread_t thread_id[MAX_THREAD]; /* POSIX Thread IDs */pthread_attr_t attr; /* Thread attributes NULL=use default */ pthread_mutex_t my_mutex; /* MUTEX data structure */#define MAX_SIZE 4000000double array[MAX_SIZE]; /* What we are summing... */void hpcwall(double *); main() {int i,retval; pthread_t tid;double single,multi,begtime,endtime;/* Initialize things */ for (i=0; i<MAX_SIZE; i++) array[i] = drand48();pthread_attr_init(&attr); /* Initialize attr with defaults */ pthread_mutex_init (&my_mutex, NULL); pthread_attr_setscope(&attr, PTHREAD_SCOPE_SYSTEM); /* Single threaded sum */GlobSum = 0; hpcwall(&begtime); for(i=0; i<MAX_SIZE;i++) GlobSum = GlobSum + array[i];hpcwall(&endtime); single = endtime - begtime;printf("Single sum=%lf time=%lf\n",GlobSum,single);/* Use different numbers of threads to accomplish the same thing */ for(ThreadCount=2;ThreadCount<=MAX_THREAD; ThreadCount++) { printf("Threads=%d\n",ThreadCount);GlobSum = 0; hpcwall(&begtime); for(i=0;i<ThreadCount;i++) { index[i]= i; retval = pthread_create(&tid,&attr,SumFunc,(void *) index[i]);thread_id[i] = tid;} for(i=0;i<ThreadCount;i++) retval = pthread_join(thread_id[i],NULL);hpcwall(&endtime); multi = endtime - begtime;printf("Sum=%lf time=%lf\n",GlobSum,multi); printf("Efficiency = %lf\n",single/(multi*ThreadCount));} /* End of the ThreadCount loop */ }void *SumFunc(void *parm){ int i,me,chunk,start,end;double LocSum; /* Decide which iterations belong to me */me = (int) parm; chunk = MAX_SIZE / ThreadCount;start = me * chunk; end = start + chunk; /* C-Style - actual element + 1 */if ( me == (ThreadCount-1) ) end = MAX_SIZE; printf("SumFunc me=%d start=%d end=%d\n",me,start,end);/* Compute sum of our subset*/ LocSum = 0;for(i=start;i<end;i++ ) LocSum = LocSum + array[i];/* Update the global sum and return to the waiting join */pthread_mutex_lock (&my_mutex); GlobSum = GlobSum + LocSum;pthread_mutex_unlock (&my_mutex); }

First, the code performs the sum using a single thread using a for-loop. Then for each of the parallel sums, it creates the appropriate number of threads that call SumFunc( ) . Each thread starts in SumFunc( ) and initially chooses an area to operation in the shared array. The “strip” is chosen by dividing the overall array up evenly among the threads with the last thread getting a few extra if the division has a remainder.

Then, each thread independently performs the sum on its area. When a thread has finished its computation, it uses a mutex to update the global sum variable with its contribution to the global sum:


recs % addup Single sum=7999998000000.000000 time=0.256624Threads=2 SumFunc me=0 start=0 end=2000000SumFunc me=1 start=2000000 end=4000000 Sum=7999998000000.000000 time=0.133530Efficiency = 0.960923 Threads=3SumFunc me=0 start=0 end=1333333SumFunc me=1 start=1333333 end=2666666 SumFunc me=2 start=2666666 end=4000000Sum=7999998000000.000000 time=0.091018 Efficiency = 0.939829Threads=4 SumFunc me=0 start=0 end=1000000SumFunc me=1 start=1000000 end=2000000 SumFunc me=2 start=2000000 end=3000000SumFunc me=3 start=3000000 end=4000000 Sum=7999998000000.000000 time=0.107473Efficiency = 0.596950 recs %

There are some interesting patterns. Before you interpret the patterns, you must know that this system is a three-processor Sun Enterprise 3000. Note that as we go from one to two threads, the time is reduced to one-half. That is a good result given how much it costs for that extra CPU. We characterize how well the additional resources have been used by computing an efficiency factor that should be 1.0. This is computed by multiplying the wall time by the number of threads. Then the time it takes on a single processor is divided by this number. If you are using the extra processors well, this evaluates to 1.0. If the extra processors are used pretty well, this would be about 0.9. If you had two threads, and the computation did not speed up at all, you would get 0.5.

At two and three threads, wall time is dropping, and the efficiency is well over 0.9. However, at four threads, the wall time increases, and our efficiency drops very dramatically. This is because we now have more threads than processors. Even though we have four threads that could execute, they must be time-sliced between three processors. It is important to match the number of runnable threads to the available resources. In compute code, when there are more threads than available processors, the threads compete among themselves, causing unnecessary overhead and reducing the efficiency of your computation. This is even worse that it might seem. As threads are switched, they move from processor to processor and their caches must also move from processor to processor, further slowing performance. This cache-thrashing effect is not too apparent in this example because the data structure is so large, most memory references are not to values previously in cache.

It’s important to note that because of the nature of floating-point (see [link] ), the parallel sum may not be the same as the serial sum. To perform a summation in parallel, you must be willing to tolerate these slight variations in your results.

Questions & Answers

calculate molarity of NaOH solution when 25.0ml of NaOH titrated with 27.2ml of 0.2m H2SO4
Gasin Reply
what's Thermochemistry
rhoda Reply
the study of the heat energy which is associated with chemical reactions
Kaddija
How was CH4 and o2 was able to produce (Co2)and (H2o
Edafe Reply
explain please
Victory
First twenty elements with their valences
Martine Reply
what is chemistry
asue Reply
what is atom
asue
what is the best way to define periodic table for jamb
Damilola Reply
what is the change of matter from one state to another
Elijah Reply
what is isolation of organic compounds
IKyernum Reply
what is atomic radius
ThankGod Reply
Read Chapter 6, section 5
Dr
Read Chapter 6, section 5
Kareem
Atomic radius is the radius of the atom and is also called the orbital radius
Kareem
atomic radius is the distance between the nucleus of an atom and its valence shell
Amos
Read Chapter 6, section 5
paulino
Bohr's model of the theory atom
Ayom Reply
is there a question?
Dr
when a gas is compressed why it becomes hot?
ATOMIC
It has no oxygen then
Goldyei
read the chapter on thermochemistry...the sections on "PV" work and the First Law of Thermodynamics should help..
Dr
Which element react with water
Mukthar Reply
Mgo
Ibeh
an increase in the pressure of a gas results in the decrease of its
Valentina Reply
definition of the periodic table
Cosmos Reply
What is the lkenes
Da Reply
what were atoms composed of?
Moses Reply
Got questions? Join the online conversation and get instant answers!
Jobilize.com Reply

Get Jobilize Job Search Mobile App in your pocket Now!

Get it on Google Play Download on the App Store Now




Source:  OpenStax, High performance computing. OpenStax CNX. Aug 25, 2010 Download for free at http://cnx.org/content/col11136/1.5
Google Play and the Google Play logo are trademarks of Google Inc.

Notification Switch

Would you like to follow the 'High performance computing' conversation and receive update notifications?

Ask