Welcome to MilkyWay@home

Posts by Crunch3r

41) Message boards : News : bad news (Message 47189)
Posted 9 Apr 2011 by Profile Crunch3r
Post:

I think we did upgrade the RAM not too long ago. But yes, a shiny new server with some SSDs would be lovely. When I bitch everyone out monday (for breaking everything and letting me know at 2am this tuesday), i'll be sure to bring it up.

And I was just finally making some progress with DNA@Home. :P


Well, to be honest, DNA@Home is broken too ;) UoTD page is stuck, Profile pages broken, pending credit page is stuck/ not updating (here as well) ... So lot's of things todo!!!

And a linux app for DNA is first priority!!! OSX is not important boinc wise ;)
42) Message boards : News : bad news (Message 47187)
Posted 9 Apr 2011 by Profile Crunch3r
Post:
Geez, I don't think Heidi even told me about that grant. :P But I think when a grant is awarded it takes some time before we actually get the money. Chances are some of it will be used for upgrades whenever it comes in.


Well it was posted almost a month ago. I guess you better talk to her soon ;)
A resonable 5 to 6k$ would get you finally some nice server and headrome to operate MW properly.


Our big problem is really with MYSQL keeping up (especially if we upgrade the ATI application ever). I think we need some kind of distributed database.


Well if you're IO bound, a massive increase in RAM is needed or some fast SSDs.
Anyway, AFIK the whole MW project still runs on FreeBSD and some crappy old server... so .. you need a new one, that's for sure.
43) Message boards : News : bad news (Message 47184)
Posted 9 Apr 2011 by Profile Crunch3r
Post:
Geez, I don't think Heidi even told me about that grant. :P But I think when a grant is awarded it takes some time before we actually get the money. Chances are some of it will be used for upgrades whenever it comes in.


Well it was posted almost a month ago. I guess you better talk to her soon ;)
A resonable 5 to 6k$ would get you finally some nice server and headrome to operate MW properly.

EDIT:

This news might be worth a proper news announcement after the dust has settled.
44) Message boards : News : bad news (Message 47182)
Posted 9 Apr 2011 by Profile Crunch3r
Post:
Modern Magellans: New NSF Grant at Rensselaer Uses the Power of the Masses to Map Dark Matter in the Galaxy



A simulation developed with MilkyWay@Home shows the formation of several stellar streams (Sagittarius, Orphan, and GD-1) around the Milky Way. The animation represents four billion years in the Milky Way, ending at the present day. Credit: Rensselaer/Benjamin A. Willett

A simulation developed with MilkyWay@Home shows the formation of several stellar streams (Sagittarius, Orphan, and GD-1) around the Milky Way.
The animation represents four billion years in the Milky Way, ending at the present day. Credit: Rensselaer/Benjamin A. Willett

Rensselaer Polytechnic Institute astronomer Heidi Newberg is using a new grant from the National Science Foundation (NSF) to begin mapping the distribution of dark matter in our galaxy. The more than $382,000 grant will utilize the massive computing power of the international MilkyWay@Home project to help uncover the whereabouts of the elusive dark matter and provide another piece in the puzzle to map the Milky Way.

Any third grader can tell you the order of the planets from the sun, but astronomers know surprisingly little about exactly how and where the mass in our galaxy is distributed across the cosmos. One large conundrum with the effort to map the Milky Way is that much of the known mass of the galaxy is completely unaccounted for. Add up the planets, sun, stars, moons, asteroids, and other known mass and only a fraction of the mass is accounted for. Many scientists purport that the difference is made up by the presence of dark matter, which is undetectable to all modern telescopic technology. But Newberg thinks that with MilkyWay@Home, the computer may be able to accomplish what no telescope has done before – show astronomers where dark matter is likely to reside in the galaxy.

Led by researchers at Rensselaer, the MilkyWay@Home project is among the fastest distributed computing programs ever in operation. At its peak, it has run at a combined computing power of over 2 petaflops donated by a total of 93,206 people in 195 countries and counting. That is more countries than the United Nations. The combined computing power of all these personal computers, which rivals the most powerful supercomputers in the world, is being used by researchers like Newberg to uncover some important and basic things about our galaxy.

“MilkyWay@Home is allowing us to consider bigger thinking when it comes to understanding the galaxy,” said Newberg, who is professor of physics, applied physics, and astronomy at Rensselaer. “With this grant we will combine my previous research mapping tidal debris streams with the power of MilkyWay@Home to simulate the creation of these streams and begin, very importantly, to start to constrain the properties of dark matter.”

Tidal debris streams, also called stellar streams, are the scattered remains of dwarf galaxies that have been ripped apart after coming too close to a larger galaxy such as the Milky Way. Several of these glowing streams currently orbit the Milky Way as part of what is aptly called the galactic halo.

Given the gross amount of dark matter thought to be present in the galaxy, the collision of dwarf galaxy and massive Milky Way that created tidal debris streams would have included the strong involvement of dark matter. This makes study of tidal debris stream formation a good window into understanding how and where dark matter comes into play in the galaxy, according to Newberg. Know how the debris stream was formed, she said, and astronomers would start to understand where dark matter may reside.

Newberg’s previous research has successfully mapped the position and distribution of stars in several stellar streams, including the Sagittarius stream, the Orphan Stream, GD-1, and the Cetus Polar Stream, using data from the Sloan Digital Sky Survey (SDSS). With this new funding, she will use MilkyWay@Home to simulate how exactly the stars in those streams got to where they currently reside in the sky.

To simulate stream creation, she will use MilkyWay@Home to model what happens when a dwarf galaxy meets up with the Milky Way and is ripped apart and spread around the galaxy like an intergalactic paint splatter. To confirm the model, she will match it to the known distribution of stars in the streams mapped by her previous work with SDSS. The first step in this process will be to constrain the properties of an initial dwarf galaxy that would be input into the MilkyWay@Home system.

“We are currently running some simulations on MilkyWay@Home right now on a single dwarf galaxy,” Newberg said. “This is a small part of the problem, but an important step in the effort to map the overall distribution of mass in the galaxy.”

The NSF grant is expected to total $382,513 over three years.

More on Newberg’s research can be found at http://www.rpi.edu/~newbeh/.

A news release on MilkyWay@Home can be found at http://news.rpi.edu/update.do?artcenterkey=2685.
Published March 15, 2011
source: http://news.rpi.edu/update.do?artcenterkey=2842
linking: http://board.mpits.net/viewtopic.php?f=14&t=49
so how about using a few $ to buy a new server ???
45) Message boards : Number crunching : AMD and NV GPU working together? (Message 47143)
Posted 7 Apr 2011 by Profile Crunch3r
Post:
before fiddling around with a cc_config.xml, you should try to figure out why the boinc client doesn't detect the ATI card in the first place...

http://milkyway.cs.rpi.edu/milkyway/show_host_detail.php?hostid=142267

Does it show up in the messages tab when starting the client (my guess, it does not) ?
46) Message boards : News : Nvidia OpenCL updated (Message 46983)
Posted 5 Apr 2011 by Profile Crunch3r
Post:
At some point after I get the new ATI application ready things will start breaking for old stuff.

Is it already known if the new ATI application will have the same requirements as the old one i.e. will it run on all ATI GPUs listed in the GPU Requirements Thread?
Yes, it should.

except for the fact that OpenCL runs only on HD 4xxx and higher (so no more 3xxx)...
In addition, if you use any OpenCL 1.1 instruction, then only HD 5xxx and above will work...


Well, according to http://milkyway.cs.rpi.edu/milkyway/forum_thread.php?id=1477&nowrap=true#46691, he's not using opencl, which is a very good thing!!

Instead, CAL++ (or calpp) is being used,which supports 38xx cards as well ;)
47) Message boards : News : Milkyway@Home for iPhone / iOS released (Message 46966)
Posted 4 Apr 2011 by Profile Crunch3r
Post:
This is not THE big news; we are still finalizing details so I can't say anything more.

-Matthew


1. We got rid of the whole cmake crap and other rather strange libraries invonked ???

2. MW finally figured out that Windows is the top OS (98.99%) and you've put linux and osx developent on the back burner where they belong ?
48) Message boards : Cafe MilkyWay : Team Recruitment III (Message 46965)
Posted 4 Apr 2011 by Profile Crunch3r
Post:


BOINC United is a unique, international team open to all that are interested.

Combining our knowledge, having inspirational talks with our Members and
helping out the scientific community by participating in various BOINC Projects.

Be part of something special!

Are you ready to make a difference ?

Please join us



BOINC United
49) Message boards : Number crunching : Amd Hd 69xx hits 88 degree Celcius.. (Message 46671)
Posted 22 Mar 2011 by Profile Crunch3r
Post:
what is the treshold temperature for 6950? is it safe to run this card 7/24 at 80-88 celcius / 175-190 Fahrenheit? so far no problem but is there any risk?


At ~ 120°C the card will start throttling, so you're safe and don't need to worry about it.
50) Message boards : Number crunching : Linux ATI application ? (Message 46670)
Posted 22 Mar 2011 by Profile Crunch3r
Post:
did i missed something?
It's still coming. I was going to put up something to test with a few weeks ago but then I got sidetracked with lots of other things.


You're going to use CAL++ instead of OCL now (i hope so)?
51) Message boards : Number crunching : GPU Requirements [OLD] (Message 46589)
Posted 17 Mar 2011 by Profile Crunch3r
Post:


Thanks. I downloaded, unzipped and ran the program. It appears I need some instructions on how to use it. I found a txt file called stderr. Here is what is reports from my last attempted run.



1. shut down boinc completely
2. open the following folder "C:\ProgramData\BOINC\projects\milkyway.cs.rpi.edu_milkyway"
3. copy all files that you've extracted from the zip file into that folder
4. start boinc again

52) Message boards : Number crunching : Strange things on HD5970 (Message 46507)
Posted 8 Mar 2011 by Profile Crunch3r
Post:
You might want to give the new 11.4 a try.
It fixed all the GPU clock issues for me and includes a new version of the stream SDK/APP runtime 2.4...

http://bit.ly/hoPFBu
53) Message boards : Cafe MilkyWay : Team Recruitment III (Message 46337)
Posted 21 Feb 2011 by Profile Crunch3r
Post:


BOINC United is a unique, international TEAM open to all that are interested.

Combining our knowledge, having inspirational talks with our Members and
helping out the scientific community by participating in various BOINC Projects.

Be part of something special!

Are you ready to make a difference ?

Please join us



BOINC United
54) Message boards : Number crunching : Validator stuck? (Message 45808)
Posted 27 Jan 2011 by Profile Crunch3r
Post:
I bet the database problems have something to do with this....


Probably... the last stats export is dated 25-Jan-2011 00:10

Anyway, i find i hard to believe that either repairing the team table or reimporting it from the the "boinc wide teams" can take that long...

I'd guess someone's not considering it to be of any importance at all :(



Seems that something has gone on at MW.


Common sense? Yeah... i agree....

Ain't nothing new, since almost anyone with a "secondary modern school qualification" get's a PhD or Dr degree at Rensselaer Polytechnic Institute these days...

Fixing a corrupt sql table, which should be a task done in 5 min in a real world scenario, where people actually work for their money...



The credit for de_separation_23_3s is 220.xx, instead of the 214 that has been the norm for some time.

Please admin, stop workflow if we do not get credit for task.
upload/download server online...


The credit is being awarded, just the info isn't being sent out to the stat sites.

55) Message boards : Number crunching : Validator stuck? (Message 45804)
Posted 26 Jan 2011 by Profile Crunch3r
Post:
I bet the database problems have something to do with this....


Probably... the last stats export is dated 25-Jan-2011 00:10

Anyway, i find i hard to believe that either repairing the team table or reimporting it from the the "boinc wide teams" can take that long...

I'd guess someone's not considering it to be of any importance at all :(



56) Message boards : Number crunching : GTX 480 (Message 43531)
Posted 5 Nov 2010 by Profile Crunch3r
Post:
What is 'the opti app.'? I have just installed an ASUS GTX480 and all work ends up in calculation error.
Boinc shows: NVIDIA GPU 0: GeForce GTX 480 (driver version 26099, CUDA version 3020, compute capability 2.0, 1503MB, 778 GFLOPS peak)


FWIW, it's not an "optimized" app. I just a recompile of the stock .024 cuda app code with a minor change to allow processing on GT4XX cards.

I only did this, because the project "devs" were not able/unwilling to recompile the app to add the support for the GT4XX cards...

It's just a few lines of code that need to be modded...



int choose_cuda_13()
{
  //check for and find a CUDA 1.3 (double precision)
  //capable card
  int device_count;
  cutilSafeCall(cudaGetDeviceCount(&device_count));
  fprintf(stderr, "Found %d CUDA cards\n", device_count);
  if (device_count < 1)
    {
      fprintf(stderr, "No CUDA cards found, you cannot run the GPU version\n");
      exit(1);
    }
  int *eligable_devices = (int*)malloc(sizeof(int) * device_count);
  int eligable_device_idx = 0;
  int max_gflops = 0;
  int device;
  char *chosen_device = 0;
  for(int idx = 0;idx<device_count;++idx)
    {
      cudaDeviceProp deviceProp;
      cutilSafeCall(cudaGetDeviceProperties(&deviceProp, idx));
      fprintf(stderr, "Found a %s\n", deviceProp.name);
      if ((deviceProp.major == 1 && deviceProp.minor == 3) || deviceProp.major == 2)
	{
	  eligable_devices[eligable_device_idx++] = idx;
	  fprintf(stderr, "Device can be used it has double precision support\n");
	  //check how many gflops it has
	  int gflops = deviceProp.multiProcessorCount * deviceProp.clockRate;
	  if (gflops >= max_gflops)
	    {
	      max_gflops = gflops;
	      device = idx;
	      if (chosen_device)
		free(chosen_device);
	      chosen_device = (char*) malloc(sizeof(char) * strlen(deviceProp.name)+1);
	      strncpy(chosen_device, deviceProp.name, strlen(deviceProp.name));
	      chosen_device[strlen(deviceProp.name)] = '\0';
	    }
	}
      else
	{
	  fprintf(stderr, "Device cannot be used, it does not have double precision support\n");
	}
    }
  free(eligable_devices);
  if (eligable_device_idx < 1) {
    fprintf(stderr, "No compute capability 1.3 or 2.x cards have been found, exiting...\n");
    free(chosen_device);
    return -1;
  } else {
    fprintf(stderr, "Chose device %s\n", chosen_device);
    cutilSafeCall(cudaSetDevice(device));  
    free(chosen_device);
    return device;
  }
}


Anyway, from looking at the code of the 0.24 app (and especially it's somewhat faulty successor v0.25).. there's still room for improving the code dramatically... the approximation of the exp function in in 0.25 is a good start but it's quite slow that way...

There are options to approximate the exp using some modified remez algorithm...
for SSE2 capable CPUs, it looks like this...

static inline __m128d _mm_fexp_poly_pd(__m128d x1) // precise,but slower than _mm_exp_pd
{
/*remez11_0_log2_sse*/
__m128i k1;
__m128d p1,a1;
__m128d xmm0, xmm1;

const __m128i offset = _mm_setr_epi32(1023, 1023, 0, 0);
x1 = _mm_min_pd(x1, _mm_set1_pd( 129.00000));
x1 = _mm_max_pd(x1, _mm_set1_pd(-126.99999));

        xmm0 = _mm_load_pd(log2e);
        xmm1 = _mm_setzero_pd();
        a1 = _mm_mul_pd(x1, xmm0);
        /* k = (int)floor(a); p = (float)k; */
        p1 = _mm_cmplt_pd(a1, xmm1);
        p1 = _mm_and_pd(p1, DONE);
        a1 = _mm_sub_pd(a1, p1);
        k1 = _mm_cvttpd_epi32(a1); // ipart
        p1 = _mm_cvtepi32_pd(k1);
        /* x -= p * log2; */
        xmm0 = _mm_load_pd(c1);
        xmm1 = _mm_load_pd(c2);
        a1 = _mm_mul_pd(p1, xmm0);
        x1 = _mm_sub_pd(x1, a1);
        a1 = _mm_mul_pd(p1, xmm1);
        x1 = _mm_sub_pd(x1, a1);
        /* Compute e^x using a polynomial approximation. */

		xmm0 = _mm_load_pd(w11);
        xmm1 = _mm_load_pd(w10);

		a1 = _mm_mul_pd(x1, xmm0);
        a1 = _mm_add_pd(a1, xmm1);
        xmm0 = _mm_load_pd(w9);
        xmm1 = _mm_load_pd(w8);
        a1 = _mm_mul_pd(a1, x1);
        a1 = _mm_add_pd(a1, xmm0);
        a1 = _mm_mul_pd(a1, x1);
        a1 = _mm_add_pd(a1, xmm1);

		xmm0 = _mm_load_pd(w7);
        xmm1 = _mm_load_pd(w6);
        a1 = _mm_mul_pd(a1, x1);
        a1 = _mm_add_pd(a1, xmm0);
        a1 = _mm_mul_pd(a1, x1);
        a1 = _mm_add_pd(a1, xmm1);

        xmm0 = _mm_load_pd(w5);
        xmm1 = _mm_load_pd(w4);
        a1 = _mm_mul_pd(a1, x1);
        a1 = _mm_add_pd(a1, xmm0);
        a1 = _mm_mul_pd(a1, x1);
        a1 = _mm_add_pd(a1, xmm1);

        xmm0 = _mm_load_pd(w3);
        xmm1 = _mm_load_pd(w2);
        a1 = _mm_mul_pd(a1, x1);
        a1 = _mm_add_pd(a1, xmm0);
        a1 = _mm_mul_pd(a1, x1);
        a1 = _mm_add_pd(a1, xmm1);

        xmm0 = _mm_load_pd(w1);
        xmm1 = _mm_load_pd(w0);
        a1 = _mm_mul_pd(a1, x1);
        a1 = _mm_add_pd(a1, xmm0);
        a1 = _mm_mul_pd(a1, x1);
        a1 = _mm_add_pd(a1, xmm1);
        /* p = 2^k; */
        k1 = _mm_add_epi32(k1, offset);
        k1 = _mm_slli_epi32(k1, 20);
        k1 = _mm_shuffle_epi32(k1, _MM_SHUFFLE(1,3,0,2));
        p1 = _mm_castsi128_pd(k1);
        /* a *= 2^k. */
        return  _mm_mul_pd(a1, p1);

}
  


This can be modified for cuda as well...
Anyway... I might do that when i have the time to do so, or modify Gipsels approach using a LUT, which is quite faster..Using my implementation of it, it beats Gipsels CPU SSE3 version by 20% in performance but it's not precise enough... Something i might work on when i have the time to do so...

Basically it's just fun beating the crap out of the stock apps... having a good laugh at the stock code and of course, waiting for the OpenCL ATI app... that'll be the icing on the cake and history will repeat itself....
57) Message boards : Number crunching : Total estimated CPU-time increases constantly during calculations - why? (Message 43530)
Posted 5 Nov 2010 by Profile Crunch3r
Post:
You got it .... its the source of much frustration, however, structuring the code this way is the only real alternative if the original design philosophy of complete software independence from any other factor is to be achieved. Many argue that its too much to chew on and the original design criteria is too ambitious. I'll leave the resolution of that one to the gurus ... greater minds than mine as they say rofl :)

Regards
Zy


The only solution to 'fix' such issues is to rewrite the boinc client code to match your needs, as the devs seem to be kind of resistant to input regarding such matters...

I pity all of you that run stock boinc clients...

LOL...
58) Message boards : Number crunching : What is current application (Message 43377)
Posted 1 Nov 2010 by Profile Crunch3r
Post:
Please forgive a simple question from a set-and-forget user.

I see a lot of talk here about 0.21 applications, and nbody applications. I appear to be running MilkyWay 0.19 only. Is there any benefit to getting later versions, and if so, how is it done?


No need to change anything. In fact the "new" apps are worse and like 5 times slower than what they should be.

The slowdown is basically because the usage of a crappy compiler and developing the new app on a OS that only 5% of all attached machines are running, instead of developing it on windows and using a proper compiler...


I think in english, such practice is called "shooting yourself in the foot"...

59) Message boards : News : a fix for the output file issue (Message 43097)
Posted 22 Oct 2010 by Profile Crunch3r
Post:
Is there room for more opt. of the new CPU app?


In general, the 0.4s should be ~30% faster than before (and it is on Linux / OS X). It's only Windows where the 0.4s are very, very slow (over 200% slower than on the same hardware in Linux). I've figured out the main reason through, and have a temporary fix which hasn't been put on the servers yet. It turns out the combination of standard math library functions from the MSVCRT and the ones replaced by MinGW are pretty terrible. Building the separation with crlibm (which should actually be slower, and it is in Linux/OS X, since crlibm has slower, more precise math functions) actually ends up being much faster on Windows. Built with crlibm, it runs only 33% slower than in Linux, so about the same speed that the old ones would be. I have to come up with a better solution with faster / less precise math on Windows to get closer to Linux.


Interesting... How about not using minGW at all for the windows platform ?
VS express doesn't cost a dime...


Why the GPUs didn't need a new app?



They do. We do not have the source for the ATI application, which is a major problem so I'm working on replacing it.


Hopefully that will NOT be using openCL... you should take a look at this one -> http://sourceforge.net/projects/calpp/
60) Message boards : Number crunching : Wow! CPU WUs done in just 160 secs! (Message 42975)
Posted 19 Oct 2010 by Profile Crunch3r
Post:
Yeah, his account shows 0 credits now. Only thing left todo is remove the equal amount (27,635,131.00) from his team credits as well.


Already done. :)


Just noticed it. Thanks.


Previous 20 · Next 20

©2024 Astroinformatics Group