Message boards :
Number crunching :
How to reduce VRAM usage for NVIDIA GPU tasks?
Message board moderation
Author | Message |
---|---|
Send message Joined: 29 Jul 14 Posts: 19 Credit: 3,451,802,406 RAC: 0 |
So I have a TITAN V that I want to use on this project while it's not doing anything important. The problem is I can't max out its usage by running more WUs simultaneously because I max out the VRAM on the TITAN and all of the work units end in 'computation error'. I can run about 8 or so WUs simultaneously if I micromanage them so they don't hit 12GB VRAM usage, but surely there's a way to set the WUs to use less VRAM somehow right? |
Send message Joined: 2 Oct 16 Posts: 167 Credit: 1,007,404,252 RAC: 29,402 |
Don't run 8x... I am running 4x with 330mb of memory usage. Is it so much higher on NV cards? Is that really 70-90 seconds at 8x? |
Send message Joined: 29 Jul 14 Posts: 19 Credit: 3,451,802,406 RAC: 0 |
Yeah look trust me, I'd need probably 16 WUs simultaneously to saturate the TITAN V's FP64. I know on 280X's the VRAM usage is significantly lower, for some reason on the TITAN, each WU uses about 1.5GB of VRAM. I'm not sure if this is because of the new architecture or if it's just an NVIDIA thing. The WUs process in about 55 - 65 seconds even with 10 WUs running simultaneously, and that's still peaking at only 70 - 75% usage, indicating there's still headroom left. |
Send message Joined: 2 Oct 16 Posts: 167 Credit: 1,007,404,252 RAC: 29,402 |
That's ridiculous. They probably complete so fast to keep it busy. A longer task, the individual piece of the bundle, would probably suit that very well. They it could probably stay busy with fewer tasks. I don't run MW on any of my Pascal/Maxwell cards cause... well they suck at FP64 ha. I' guess it would be NV's implementation of OpenCL code vs one NV card over another. A guess. |
Send message Joined: 24 Jan 11 Posts: 712 Credit: 553,745,311 RAC: 61,771 |
Generally most current OpenCL applications are limited to 25% of VRAM on graphics cards. So you only have approximately 3GB of the 12GB of VRAM accessible on your Titan V for MW tasks to use. If and when applications start using the OpenCL 2.0 specification that opens up global_work_size memory space, then you would be able to fully access the 12GB. From the Nvidia driver release notes.
|
Send message Joined: 29 Jul 14 Posts: 19 Credit: 3,451,802,406 RAC: 0 |
Well maybe that's how it's set out in NVIDIA's guidelines, but Milkyway still allocates up to 12GB of VRAM. Here's a graph from HWiNFO64 that shows my VRAM usage with 8 WUs running simultaneously, clearly showing it's above the 3GB threshold (Y-axis is from 0 to 12500MB of VRAM allocation). https://i.imgur.com/665amcH.png |
Send message Joined: 2 Oct 16 Posts: 167 Credit: 1,007,404,252 RAC: 29,402 |
Maybe 3gb per task then. Otherwise why put 4+ GB on any AMD card. |
Send message Joined: 8 May 09 Posts: 3339 Credit: 524,010,781 RAC: 1 |
Maybe 3gb per task then. Otherwise why put 4+ GB on any AMD card. Because crunching is NOT their primary market, it's gaming and they can access it all. Building super computers is a big market share too and they can access it all too. |
Send message Joined: 13 Dec 12 Posts: 101 Credit: 1,782,758,310 RAC: 0 |
A Titan V as in the $3000 Volta card Cautilus??? Jeezus. I thought I had the BOINC bug bad...LOL. |
Send message Joined: 2 Oct 16 Posts: 167 Credit: 1,007,404,252 RAC: 29,402 |
A Titan V as in the $3000 Volta card Cautilus??? Yes, x2 cards. NV put a hefty price tag on their top compute card. |
Send message Joined: 13 Dec 12 Posts: 101 Credit: 1,782,758,310 RAC: 0 |
WOW...that has a staggering DP score. 6144.0 (7449.6) Crazy! (Guess it comes with a crazy price tag too) Cautilus couldn't do any better to try and catch Gary Roberts ;) |
©2024 Astroinformatics Group