Welcome to MilkyWay@home

updating the stock osx/linux/windows applications

Message boards : News : updating the stock osx/linux/windows applications
Message board moderation

To post messages, you must log in.

AuthorMessage
Profile Travis
Volunteer moderator
Project administrator
Project developer
Project tester
Project scientist

Send message
Joined: 30 Aug 07
Posts: 2046
Credit: 26,480
RAC: 0
Message 40408 - Posted: 15 Jun 2010, 18:50:13 UTC

I'm going to be updating the stock applications in the next few days because I've discovered a bug in them when running the latest searches.


I just updated the OSX applications (to version 0.30). I had to modify the checkpointing a bit, so let me know if they are checkpointing correctly.


I'll be updating linux and windows tonight.
ID: 40408 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Profile Dirk Sadowski

Send message
Joined: 30 Apr 09
Posts: 101
Credit: 29,874,293
RAC: 0
Message 40425 - Posted: 15 Jun 2010, 23:51:06 UTC

Hi,

O.K., it's not the new app for CUDA which you mentioned, but I hope it's the correct thread for to write this.

I have a GTX260-216 and BOINC DLed the stock CUDA app for MilkyWay.

BOINC got:
milkyway_0.24_windows_intelx86__cuda23.exe
cudart.dll
cutil32.dll

But the strange is that cudart.dll is version 6.14.11.2020, this mean it's V2.2 and not V2.3 .
Is there something wrong?
Or is the file only wrong marked?

At SETI@home the difference between CUDA V2.2 and V2.3 mean ~ 30 % speed difference.

Best regards.

ID: 40425 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Emanuel

Send message
Joined: 18 Nov 07
Posts: 280
Credit: 2,442,757
RAC: 0
Message 40437 - Posted: 16 Jun 2010, 11:36:23 UTC - in response to Message 40425.  
Last modified: 16 Jun 2010, 11:36:57 UTC

Somewhat related, the just released 257.21 GeForce driver release notes mention that the driver "Adds support for CUDA Toolkit 3.1 which includes significant performance increases for double precision math operations." I imagine this probably refers to new instructions/functions rather than a significant speed-up for existing ones, unless they decided to remove the double precision operations cap for the 400 series cards. Even so, it might be worth checking out.
ID: 40437 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Profile Dirk Sadowski

Send message
Joined: 30 Apr 09
Posts: 101
Credit: 29,874,293
RAC: 0
Message 40440 - Posted: 16 Jun 2010, 16:35:57 UTC
Last modified: 16 Jun 2010, 16:47:48 UTC

BTW.

The current available optimized CPU app have a version number of 0.21 .

This 'old' optimized CPU app will calculate well the 'new' WUs (after mod of the app_info.xml file) and will match well with the results of the new apps (wingman)?
ID: 40440 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
MacRonin

Send message
Joined: 23 Jan 10
Posts: 4
Credit: 172,174
RAC: 0
Message 40491 - Posted: 17 Jun 2010, 17:43:13 UTC

Not sure if its' related to the upgrade to .30 but I am running.30 on Mac OSX 105.8 and I am getting a strange result.

According to the properties display the checkpoints are being taken, but the percentage complete has been stuck at 0.0% since it started. So far two WU's have over 11 & 13 hours of CPU time

http://milkyway.cs.rpi.edu/milkyway/result.php?resultid=143137141
http://milkyway.cs.rpi.edu/milkyway/result.php?resultid=143137142

I have tried rebooting the system, but that had no impact on the numbers.

While it is currently in "waiting to run" status (probably because is completion deadline is 20 minutes later and the others are running as high priority). the WU ( http://milkyway.cs.rpi.edu/milkyway/result.php?resultid=142461244 ) does have a progressing completion percentage.
ID: 40491 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Profile GPV67
Avatar

Send message
Joined: 12 Jan 08
Posts: 5
Credit: 3,994,193
RAC: 0
Message 40492 - Posted: 17 Jun 2010, 18:10:58 UTC - in response to Message 40491.  

Hi,
It seems there is an issue with checkpointing on Mac OS X.
Percentage complete has been stuck at 0.0% since tasks are started (18 hours)
As my MacPro is running 12 h a day, WU is running since 2 days.
below copy of astronomy_checkpoint file of same WU before and after machine restart

After 6 hours of machine uptime:
background_integral: 0.00000000000000000000
stream_integrals[3]: 0.00000000000000000000, 0.00000000000000000000, 0.00000000000000000000
prob_sum: 0.00000000000000000000, num_zero: 0, bad_jacobians: 0
current_star_point: 0
current_integral: 0
number_integrals: 1
mu[min,max,steps]: 135.000, 240.000, 1600
nu[min,max,steps]: -1.250, 1.250, 640
r[min,max,steps]: 16.000, 22.500, 1400
mu_step: 1317, nu_step: 107, r_step: 0
background_integral: 0.00231811528309438597
stream_integrals[3]: 1259.58644129462277305720, 967.01620570881334515434, 3198.81638506128138033091

After few minutes of machine uptime:
background_integral: 0.00000000000000000000
stream_integrals[3]: 0.00000000000000000000, 0.00000000000000000000, 0.00000000000000000000
prob_sum: 0.00000000000000000000, num_zero: 0, bad_jacobians: 0
current_star_point: 0
current_integral: 0
number_integrals: 1
mu[min,max,steps]: 135.000, 240.000, 1600
nu[min,max,steps]: -1.250, 1.250, 640
r[min,max,steps]: 16.000, 22.500, 1400
mu_step: 4, nu_step: 105, r_step: 0
background_integral: 0.00231991176126485499
stream_integrals[3]: 1263.15242429980276028800, 967.01718327095272798033, 3200.90837238598305702908
ID: 40492 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Bill Barnsley

Send message
Joined: 5 Apr 09
Posts: 1
Credit: 517,002
RAC: 0
Message 40495 - Posted: 17 Jun 2010, 21:05:02 UTC - in response to Message 40408.  

Nope - checkpointing is not working on OSX. The work unit has been working for 15 hours, shows 10 hours to go, and shows 0% progress.
ID: 40495 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Profile arkayn
Avatar

Send message
Joined: 14 Feb 09
Posts: 999
Credit: 74,932,619
RAC: 0
Message 40496 - Posted: 17 Jun 2010, 22:01:50 UTC

I looked into my WU this afternoon and can report the exact same thing as everybody else. It just sits there and does not actually progress.

Looks like time for 0.31.
ID: 40496 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Profile Travis
Volunteer moderator
Project administrator
Project developer
Project tester
Project scientist

Send message
Joined: 30 Aug 07
Posts: 2046
Credit: 26,480
RAC: 0
Message 40497 - Posted: 17 Jun 2010, 22:23:16 UTC - in response to Message 40496.  

I found the problem, I should have the OS X applications updated tonight.

-Travis
ID: 40497 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
MacRonin

Send message
Joined: 23 Jan 10
Posts: 4
Credit: 172,174
RAC: 0
Message 40498 - Posted: 17 Jun 2010, 22:39:34 UTC - in response to Message 40497.  

Should we cancel?
or is it actually dong valid work but not using the checkpoint/% complete functions correctly?
ID: 40498 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Profile arkayn
Avatar

Send message
Joined: 14 Feb 09
Posts: 999
Credit: 74,932,619
RAC: 0
Message 40499 - Posted: 17 Jun 2010, 22:46:23 UTC

I just suspended my 7 units and will install the new app when it comes out.
ID: 40499 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Henry Bundy

Send message
Joined: 2 Mar 10
Posts: 5
Credit: 105,634,798
RAC: 0
Message 40523 - Posted: 18 Jun 2010, 17:55:04 UTC

V .30 for Mac is terrible. The Boink manager "Progress" and the "Time to Completion" don't work.
ID: 40523 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Profile Travis
Volunteer moderator
Project administrator
Project developer
Project tester
Project scientist

Send message
Joined: 30 Aug 07
Posts: 2046
Credit: 26,480
RAC: 0
Message 40534 - Posted: 19 Jun 2010, 0:24:03 UTC - in response to Message 40523.  

V .30 for Mac is terrible. The Boink manager "Progress" and the "Time to Completion" don't work.


Should be fixed now in version 0.31
ID: 40534 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Profile The Gas Giant
Avatar

Send message
Joined: 24 Dec 07
Posts: 1947
Credit: 240,884,648
RAC: 0
Message 40536 - Posted: 19 Jun 2010, 1:26:13 UTC

Does CP's op ap need updating as well?
ID: 40536 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Profile Travis
Volunteer moderator
Project administrator
Project developer
Project tester
Project scientist

Send message
Joined: 30 Aug 07
Posts: 2046
Credit: 26,480
RAC: 0
Message 40538 - Posted: 19 Jun 2010, 1:43:54 UTC - in response to Message 40536.  

Does CP's op ap need updating as well?


Yeah, there's a bunch of updates that we need to get done so we can swap over to the assimilator/validator that doesn't generate files for each WU and for each result.

Hopefully he gets settled at his new post doc and can work with us or send us the code so we can update it.
ID: 40538 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Cluster Physik

Send message
Joined: 26 Jul 08
Posts: 627
Credit: 94,940,203
RAC: 0
Message 40556 - Posted: 20 Jun 2010, 16:38:28 UTC - in response to Message 40538.  

Does CP's op ap need updating as well?

Yeah, there's a bunch of updates that we need to get done so we can swap over to the assimilator/validator that doesn't generate files for each WU and for each result.

Hopefully he gets settled at his new post doc and can work with us or send us the code so we can update it.

I will start my post doc position in two weeks.

Has anything changed since this post here and your summary there?
That version from april was already compatible with parameters given in the command line and wrote the integrals and likelihood also separately to the stderr.txt. Changes to the likelihood calculation should be really easy, as it is done on the CPU, not the GPU.

I will try to update the ATI app to use the newer Stream SDK (I was still using the outdated SDK 1.3 so far) this evening. Some changes will be needed, as SDK 1.3 didn't support multiple GPUs or the CPU release during the GPU kernels, I implemented that as some kind of small hack to the old SDK (it is open source). But the newer SDKs have official support for those features (that's why I want to use it), but it works of course a bit different than my version. I will see if all fits nicely or if some more changes are needed.

I've done also some changes to the boinc_api to fix some problems there, I don't know if it is still necessary or if Berkeley has fixed those issues (like the suspend/resume problem when using critical sections) since the last time I looked at it. But I will detail it when I send the code.

And be warned already, I never had time to clean it up, so it may look a bit convoluted. After all, it still bases on the 0.07 code, I only adapted it to stay compatible with the newer versions without investing any effort to increase the readability.
ID: 40556 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Cluster Physik

Send message
Joined: 26 Jul 08
Posts: 627
Credit: 94,940,203
RAC: 0
Message 40561 - Posted: 21 Jun 2010, 0:57:06 UTC
Last modified: 21 Jun 2010, 1:13:12 UTC

Some update from my side:

It runs now with the newer SDK version, but unfortunately I had to recognize, that a lot of nasty bugs or "features" survived. Heck, I complained about wrong indexing when using double precision constant buffers 18 months ago, I even sent a test case to ATI's stream developer team back then! That makes it more or less impossible to write the ATI GPU code completely in a high level language, one needs to correct the code on the underlying CAL/IL level. I can only hope that the OpenCL compiler was more in the focus of ATI lately. It really looks like nobody actually tested the double precision stuff within their compilers as there are plenty of problems with it but generally it works with single precision.

@Travis:
I hope you don't want to rewrite the integration kernels from scratch (I would almost suggest looking at OpenCL if that need arises, or alternatively CAL++ for ATIs, if you want more speed). But as I said before, changes to the likelihood stuff are no problem at all.

Edit:
I've just seen you added the possibility to use another "background profile" in the MW3 0.05 code (besides quite some other changes). If used, one would need to rewrite the GPU kernels. I will see, if I can look into it tomorrow.
ID: 40561 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Profile Travis
Volunteer moderator
Project administrator
Project developer
Project tester
Project scientist

Send message
Joined: 30 Aug 07
Posts: 2046
Credit: 26,480
RAC: 0
Message 40564 - Posted: 21 Jun 2010, 2:42:02 UTC - in response to Message 40561.  

Actually we've been trying to port everything into openCL. We've moved the code over to GitHub, and Matt Arsenault has been working on making a nice build system (in cmake) which should make changes and compiling much simpler.

http://github.com/arsenm/milkywayathome_client

The changes are pretty simple, I think the only GPU kernel that would need to be modified is the likelihood one. If you look at evaluation_optimized.c:

http://github.com/arsenm/milkywayathome_client/blob/master/milkyway_separation/cpu/evaluation_optimized.c

The big changes are that we're also calculating separate likelihoods for each stream and the background, lines 680-710 or so, in addition to full likelihood. We're also using kahan summation in the CPU code too, so I'm hoping that will bring the values closer.

Matt's also been working on a test system, so after cmake has made the makefiles to compile, it can also automatedly run the set of tests we have so far. I'm gonna direct Matt here because he knows more about it.

So really the only big change is calculating the stream and background likelihoods separately in addition to the full likelihood. The input to the program is a little different as well (all through the command line and we can specify the name of the input star and parameter file there).

It would be really great if we could get the ATI code merged into the code on github (I think we can do this via OpenCL) so there's one nice place for everyone to work on it.
ID: 40564 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Matt Arsenault
Volunteer moderator
Project developer
Project tester
Project scientist

Send message
Joined: 8 May 10
Posts: 576
Credit: 15,979,383
RAC: 0
Message 40565 - Posted: 21 Jun 2010, 3:32:32 UTC - in response to Message 40564.  

Actually we've been trying to port everything into openCL.

I think OpenCL is really the best option at this point. I'm hoping to work on fixing it later in the summer if I get a chance. I'm working mostly now on getting the N-body stuff into a useful state, and I have lots of reading to do on OpenCL first. It would be great if you want to help work on that.


We've moved the code over to GitHub, and Matt Arsenault has been working on making a nice build system (in cmake) which should make changes and compiling much simpler.

Well, it is a major improvement over the old giant Makefile, but it isn't exactly as nice as it should be yet.

Matt's also been working on a test system, so after cmake has made the makefiles to compile, it can also automatedly run the set of tests we have so far. I'm gonna direct Matt here because he knows more about it.

The tests right now are only semi-useful, but eventually it will be nice.


ID: 40565 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote

Message boards : News : updating the stock osx/linux/windows applications

©2024 Astroinformatics Group