ldp5500
|
|
October 16, 2015, 03:12:39 AM |
|
Is anyone else having issues with random closes while mining quark with the latest .70? I have 2 different gtx 970 rigs, windows 10. Both computers will crash within the hour mining at completely stock clocks. Can someone post a quark bat file so I can see how yours are setup.
TYIA.
:loop ccminer -a quark -o stratum+tcp:\\poolspecificaddress:port -u username.worker1 -p pswd goto loop ccminer -a quark -o stratum+tcp:\\poolspecificaddress:port -u username.worker1 -p pswd pause
|
|
|
|
rekphiv
Newbie
Offline
Activity: 2
Merit: 0
|
|
October 16, 2015, 04:39:14 AM |
|
Thanks, for some reason my Quark file did not have the loop setup. Appears to be working correctly after crash now, Will test tonight.
|
|
|
|
Slava_K
|
|
October 16, 2015, 05:09:11 AM |
|
Latest commit speed degrade on lyra2rev2 750Ti 15kH, 980GTX 20kH.
|
|
|
|
dominuspro
|
|
October 16, 2015, 05:24:42 AM |
|
Is anyone else having issues with random closes while mining quark with the latest .70? I have 2 different gtx 970 rigs, windows 10. Both computers will crash within the hour mining at completely stock clocks. Can someone post a quark bat file so I can see how yours are setup.
TYIA.
I have the same random crashes on 1 of 2 machines. Both running windows7x64. The loop works as a workaround...
|
|
|
|
Genoil
|
|
October 16, 2015, 05:39:34 AM |
|
wow thats actually bad news because it seems memory heavy algos are the future for gpus As this is a (proven) Windows driver issue, it's down to the driver guys to fix it. But something tells they aren't too hasty until the game devs start complaining . Too bad GTX' cant be put in TCC mode..
|
ETH: 0xeb9310b185455f863f526dab3d245809f6854b4d BTC: 1Nu2fMCEBjmnLzqb8qUJpKgq5RoEWFhNcW
|
|
|
hashbrown9000
|
|
October 16, 2015, 07:16:33 AM |
|
Anyone know what the long-form entry is for these switches: "-L" , "-l" ? Upper and lowercase - "el". I run ccminer with a .conf file so I need to write out the switch names. i.e. becomes
|
Pinkcoin: ETH: VTC: BTC:
|
|
|
sp_ (OP)
Legendary
Offline
Activity: 2912
Merit: 1087
Team Black developer
|
|
October 16, 2015, 07:22:01 AM |
|
-d, --devices Comma separated list of CUDA devices to use. \n\ Device IDs start counting from 0! Alternatively takes\n\ string names of your cards like gtx780ti or gt640#2\n\ (matching 2nd gt640 in the PC)\n\ -i --intensity=N GPU intensity 8-31 (default: auto) \n\ Decimals are allowed for fine tuning \n\ -f, --diff Divide difficulty by this factor (std is 1) \n\ -v, --vote=VOTE block reward vote (for HeavyCoin)\n\ -m, --trust-pool trust the max block reward vote (maxvote) sent by the pool\n\ -o, --url=URL URL of mining server\n\ -O, --userpass=U:P username:password pair for mining server\n\ -u, --user=USERNAME username for mining server\n\ -p, --pass=PASSWORD password for mining server\n\ --cert=FILE certificate for mining server using SSL\n\ -x, --proxy=[PROTOCOL://]HOST[:PORT] connect through a proxy\n\ -t, --threads=N number of miner threads (default: number of nVidia GPUs)\n\ -g, --gputhreads=N number of threads per gpu (default: 1)\n\ -r, --retries=N number of times to retry if a network call fails\n\ (default: retry indefinitely)\n\ -R, --retry-pause=N time to pause between retries, in seconds (default: 30)\n\ --time-limit maximum time [s] to mine before exiting the program.\n\ -T, --timeout=N network timeout, in seconds (default: 270)\n\ -s, --scantime=N upper bound on time spent scanning current work when\n\ long polling is unavailable, in seconds (default: 5)\n\ -N, --statsavg number of samples used to display hashrate (default: 30)\n\ --no-gbt disable getblocktemplate support (height check in solo)\n\ --no-longpoll disable X-Long-Polling support\n\ --no-stratum disable X-Stratum support\n\ -q, --quiet disable per-thread hashmeter output\n\ --no-color disable colored output\n\ -D, --debug enable debug output\n\ -P, --protocol-dump verbose dump of protocol-level activities\n\ --cpu-affinity set process affinity to cpu core(s), mask 0x3 for cores 0 and 1\n\ --cpu-priority set process priority (default: 0 idle, 2 normal to 5 highest)\n\ -b, --api-bind IP/Port for the miner API (default: 127.0.0.1:4068)\n\ -S, --syslog use system log for output messages\n\ --syslog - prefix = ... allow to change syslog tool name\n\ -B, --background run the miner in the background\n\ --benchmark run in offline benchmark mode\n\ --cputest debug hashes from cpu algorithms\n\ -c, --config=FILE load a JSON-format configuration file\n\ -C, --cpu-mining Enable the cpu to aid the gpu. (warning: uses more power)\n\ -V, --version display version information and exit\n\ -h, --help display this help text and exit\n\ -X, --XIntensity intensity GPU intensity(default: auto) \n\ --broken-neo-wallet Use 84byte data for broken neoscrypt wallets.\n\ ";
|
|
|
|
hashbrown9000
|
|
October 16, 2015, 07:24:37 AM |
|
yes i checked help file but i don't see these switches listed.
|
Pinkcoin: ETH: VTC: BTC:
|
|
|
hashbrown9000
|
|
October 16, 2015, 07:41:29 AM |
|
Ok I found the switches in cudaminer help file: --launch-config [-l] specify the kernel launch configuration per device. This replaces autotune or heuristic selection. You can pass the strings "auto" or just a kernel prefix like F or K or T to autotune for a specific card generation or a kernel prefix plus a lauch configuration like F28x8 if you know what kernel runs best (from a previous autotune).
--lookup-gap [-L] values > 1 enable a tradeoff between memory savings and extra computation effort, in order to improve efficiency with high N-factor scrypt-jane coins. Defaults to 1.
|
Pinkcoin: ETH: VTC: BTC:
|
|
|
pallas
Legendary
Offline
Activity: 2716
Merit: 1094
Black Belt Developer
|
|
October 16, 2015, 07:56:37 AM |
|
Is anyone else having issues with random closes while mining quark with the latest .70? I have 2 different gtx 970 rigs, windows 10. Both computers will crash within the hour mining at completely stock clocks. Can someone post a quark bat file so I can see how yours are setup.
TYIA.
I have the same random crashes on 1 of 2 machines. Both running windows7x64. The loop works as a workaround... same happens to me on linux, so I assume it's a ccminer quark specific issue.
|
|
|
|
sp_ (OP)
Legendary
Offline
Activity: 2912
Merit: 1087
Team Black developer
|
|
October 16, 2015, 09:55:50 AM Last edit: October 16, 2015, 10:33:11 AM by sp_ |
|
2 sp_ Latest github compile is 20-30khs slower then release 70 on lyra2v2. gtx750, win7x64, cuda6.5 x32 build.
yes, I tested last night. The latest is slower on the 750/750ti. I will fix it. I have submitted 2 new checkins in the lyra2v2 algo. Can you please test? Note that the default intensity is low at -X 5, probobly bether with -X 8. on the 750ti -X 10 or 11 With the latest checkin the gtx 970 G1 windforce oc is doing 9525KHASH. on standard clock. 10+ with overclocking. The asus strix 750ti is doing 4485KHASH with -X 11 on standard clocks and 5MHASH with overclocking.
|
|
|
|
djeZo
|
|
October 16, 2015, 11:11:59 AM |
|
What speeds do you get on GTX 980 Ti and GTX 950 Lyra2REv2?
I get GTX 980 Ti ... 17.450 khs GTX 950 ... 5.480 khs
clocks? OS? build? around 1400mhs both cards, windows, latest SP... but I tweaked some params, originally I was getting 17khs on 980 Ti and 5khs on 950.
|
|
|
|
djeZo
|
|
October 16, 2015, 11:25:09 AM |
|
wow thats actually bad news because it seems memory heavy algos are the future for gpus I was experiencing same kind of issue when I was making Axiom CUDA algo. Having 980 Ti, which packs 6 gig of memory, whenever I set algo to use more than about 2,5 gigs, there was a massive slow down, bus interface load jumped up, TDP jumped down. Since 980 Ti is my primary GPU, it constantly has mem load of about 400 mega even in idle time - and that would explain that actual mem cutoff is at around 2.1 gigs - same as other v2 maxwell cards. I don't have account there to post, but measure bus interface load during these bottlenecks - maybe it can reveal another hint getting down (I used GPUZ for measuring bus interface load). Bus interface load is - to my knowledge - how much PCIE bus gets loaded with data. And my algorithm implementation was sending very very little data over this bus - not something to load PCIE 3.0 16x so massively that it would show 30-50% of load. I could not explain, why bus load was so high, googling gave no results and I kinda gave up. But now that you revealed this slow down happening with other algorithms, other cards, I have my suspicion that these problems are related. My first idea would be; what if CUDA is automatically syncing GPU and CPU memory - as if some part of GPU memory was set to be in sync with CPU memory - this would explain massive bus load, as my algo was causing a lot of changes in this massive allocated buffer. I believe, CUDA even has a name for this - Unified memory. And to my knowledge, it is only active when you explicitly set so. What if it is active even in cases when you do not explicitly set so? Or maybe a bug in CUDA software - sending data over bus even though there is no need for synced memory space?
|
|
|
|
|
pallas
Legendary
Offline
Activity: 2716
Merit: 1094
Black Belt Developer
|
|
October 16, 2015, 11:28:47 AM |
|
I was experiencing same kind of issue when I was making Axiom CUDA algo. Having 980 Ti, which packs 6 gig of memory, whenever I set algo to use more than about 2,5 gigs, there was a massive slow down, bus interface load jumped up, TDP jumped down. Since 980 Ti is my primary GPU, it constantly has mem load of about 400 mega even in idle time - and that would explain that actual mem cutoff is at around 2.1 gigs - same as other v2 maxwell cards.
I don't have account there to post, but measure bus interface load during these bottlenecks - maybe it can reveal another hint getting down (I used GPUZ for measuring bus interface load).
Bus interface load is - to my knowledge - how much PCIE bus gets loaded with data. And my algorithm implementation was sending very very little data over this bus - not something to load PCIE 3.0 16x so massively that it would show 30-50% of load. I could not explain, why bus load was so high, googling gave no results and I kinda gave up. But now that you revealed this slow down happening with other algorithms, other cards, I have my suspicion that these problems are related. My first idea would be; what if CUDA is automatically syncing GPU and CPU memory - as if some part of GPU memory was set to be in sync with CPU memory - this would explain massive bus load, as my algo was causing a lot of changes in this massive allocated buffer. I believe, CUDA even has a name for this - Unified memory. And to my knowledge, it is only active when you explicitly set so. What if it is active even in cases when you do not explicitly set so? Or maybe a bug in CUDA software - sending data over bus even though there is no need for synced memory space?
you could easily test it by running the same thing on the same card but thru a 1x raiser.
|
|
|
|
sp_ (OP)
Legendary
Offline
Activity: 2912
Merit: 1087
Team Black developer
|
|
October 16, 2015, 11:36:19 AM |
|
I was experiencing same kind of issue when I was making Axiom CUDA algo. Having 980 Ti, which packs 6 gig of memory, whenever I set algo to use more than about 2,5 gigs, there was a massive slow down, bus interface load jumped up, TDP jumped down. Since 980 Ti is my primary GPU, it constantly has mem load of about 400 mega even in idle time - and that would explain that actual mem cutoff is at around 2.1 gigs - same as other v2 maxwell cards.
I don't have account there to post, but measure bus interface load during these bottlenecks - maybe it can reveal another hint getting down (I used GPUZ for measuring bus interface load).
Bus interface load is - to my knowledge - how much PCIE bus gets loaded with data. And my algorithm implementation was sending very very little data over this bus - not something to load PCIE 3.0 16x so massively that it would show 30-50% of load. I could not explain, why bus load was so high, googling gave no results and I kinda gave up. But now that you revealed this slow down happening with other algorithms, other cards, I have my suspicion that these problems are related. My first idea would be; what if CUDA is automatically syncing GPU and CPU memory - as if some part of GPU memory was set to be in sync with CPU memory - this would explain massive bus load, as my algo was causing a lot of changes in this massive allocated buffer. I believe, CUDA even has a name for this - Unified memory. And to my knowledge, it is only active when you explicitly set so. What if it is active even in cases when you do not explicitly set so? Or maybe a bug in CUDA software - sending data over bus even though there is no need for synced memory space?
you could easily test it by running the same thing on the same card but thru a 1x raiser. Thats why I made the -g switch. You get problems in windows when allocating big buffers. Running quark with -g 2 -i 24 is using the same amount of memory as -i 25 but the 2 intensity blocks are split into two. -i 25 will cause out of memory while -g 2 -i 24 will not. But we had to add some more logic to the the g switch to work .(blake 512 rewrite) so it might be slower..
|
|
|
|
sp_ (OP)
Legendary
Offline
Activity: 2912
Merit: 1087
Team Black developer
|
|
October 16, 2015, 11:36:43 AM |
|
Thanks
|
|
|
|
Slava_K
|
|
October 16, 2015, 12:05:29 PM |
|
Lyra2REv2 -X 12 looks good (test on 750Ti, GTX960, GTX980).
|
|
|
|
sp_ (OP)
Legendary
Offline
Activity: 2912
Merit: 1087
Team Black developer
|
|
October 16, 2015, 12:34:12 PM |
|
Lyra2REv2 -X 12 looks good (test on 750Ti, GTX960, GTX980).
The default values are: gtx 980TI: -X 16 (probobly -X 24 or -X 30 is bether (-X 30 is using abit more than 3 gig of memory) gtx 980: -X 16 (probobly -X 24 or -X 30 is bether) gtx 970: -X 20 (I tried different values. -X 20 is just over 2 Gigabyte of memory and looks good) gtx 960: -X 5 (-x 10, -x 11 / -x 12) gtx 950: -X 8 (-x 10, -x 11 / -x 12) gtx 750: -X 5 (-x 8 ) gtx 750ti: -X 5 (-x 10,-x 11 / -x 12) The default -X values are probobly not optimal. I had some problems with some of my rigs (out of memory (6 cards)) (4gig system ram) So I lowered the default intensities.. For cards that trottle, use lower values for -X.
|
|
|
|
sp_ (OP)
Legendary
Offline
Activity: 2912
Merit: 1087
Team Black developer
|
|
October 16, 2015, 01:06:29 PM |
|
What speeds do you get on GTX 980 Ti and GTX 950 Lyra2REv2? I get GTX 980 Ti ... 17.450 khs GTX 950 ... 5.480 khs
clocks? OS? build? around 1400mhs both cards, windows, latest SP... but I tweaked some params, originally I was getting 17khs on 980 Ti and 5khs on 950. Would you mind to share your parameters?
|
|
|
|
|