cgminer 2.2.6 out. 2+ more mhash on 7970. It's getting harder and harder to extract much more 1200/1050+5% clocks, intensity 11 - 717 Mhash. I'm going to have to look at what methods you're using. I'm curious as to how the programming differs between VLIW and GCN. Direct link to the kernel in the git tree: https://github.com/ckolivas/cgminer/blob/master/poclbm120214.clIt's using a worksize of 256, 1 vector (i.e. no vectors) and intensity 11.
|
|
|
Hmm what do the 58xx report their name as? Cypress? It sounds like cgminer should default them to worksize 256 instead of 128. Would that be a fair assessment?
Yeah, it's Cypress. The best worksize really depends on the user correctly setting the ram clock; on the default (1GHz), it's best 64 or 128 with 4 vectors. Gonna link again the graphic I like so much (from phatk thread) Yes I remember this old graph. It would appear to me that cgminer's defaults of v2 w128 are the most robust for all tunings unless the user goes out of his way to tweak things manually, in which case he'll be choosing his own settings, thanks.
|
|
|
Hmm what do the 58xx report their name as? Cypress? It sounds like cgminer should default them to worksize 256 instead of 128. Would that be a fair assessment?
|
|
|
With Diapolo's help we finally got the diakgcn kernel working on cgminer. I've just committed code to the git tree which makes it work. Alas at the same engine and clock speeds on the 7970, diakgcn gives me 699 MHash while my customised kernel gives me 717 MHash. But now that it's working, he may be able to tweak it further...
|
|
|
Is there some way I'm not aware of (and I went back and re-read the latest docs) to set GPU core/RAM clocks on exit?
I can mine at a much higher clock/much lower memory, than I choose to game.. and if I forget to go back and reset it each time I close the miner, my desktop machine will hang..
I realize it's a small thing, but just wondering if there's a way to set it?
cgminer sets clocks all back to default on exit... if it exits cleanly, and of course on windows it's a miracle when it does.
|
|
|
But it was working perfectly on a previous version of cgminer... or maybe it was the driver changes to 11.12 and sdk 2.1 that caused this..? brb, downloading old cgminer version to find out.
Never blame the new version of cgminer. ALWAYS blame something else first. I know right? But how would the change to 11.12/2.1 cause this odd issue, which I have never had before? Any ideas on debugging this? It is acting like no changes are being made at all - voltage, core clock, memory clock, etc. Did you check what it says when you run "cgminer -n" or start it with debugging enabled as it says in the readme? --verbose -D -T as well as your regular commands. Might give a clue. Also perhaps the hardware stuff got lost for 5xxx in later ATI drivers and worked on the ones somewhere in between. Knowing AMD that's entirely possible Like it might work in 11.6 and then got broken in 11.11 or something.
|
|
|
But it was working perfectly on a previous version of cgminer... or maybe it was the driver changes to 11.12 and sdk 2.1 that caused this..? brb, downloading old cgminer version to find out.
Never blame the new version of cgminer. ALWAYS blame something else first.
|
|
|
Update on 7970 performance. I did a little calculation to see how my 7970 is performing with the latest cgminer. Clock per clock it is precisely 1/3 faster than the 6970 running at the equivalent clock speed. What's hilarious about this is that the number of stream processors is precisely 1/3 more than the 6970 (2048 vs 1536). This is despite the 6970 running on phatk on sdk 2.5 and 7970 running on my customised poclbm on sdk 2.6. As I said before, at least on the 6970 which I don't grossly underclock the memory (only -125) I'm getting slightly better performance on poclbm with 2.6 now than phatk with 2.5. So in terms of raw performance, I think the current cgminer release and kernel is doing pretty good. Alas none of the magic GCN changes that aren't just more processors appear to have been exploited, and every thread I've read on suggestions to exploit them on these forums has only led to worse performance, often disastrously worse.
|
|
|
Well I wondered that, but I thought that newer drivers refused to support older SDKs. I guess I'll try it again soon.
I don't think so but with so many combinations of hardware, OS, drivers, and SDK YMMV. Well, I got it working, but my 5870s at 950/160 are producing 370mhash now. That is suspiciously similar to stock, wonder if OC settings aren't being applied by cgminer? brb checking EDIT: When I hit G then C then 0 and enter then E and typed 950 and hit enter and then Y, cgminer told me that the driver accepted the voltages, but that I should "check the values below". It didn't seem to actually OC the cards. Right, that's why cgminer reports back the *actual* values because the driver can happily accept some changes and the card can happily ignore them.
|
|
|
I created the .bat and the shortcut to make CGMiner start up, but the guide had zero information in it about setting up a .config file, i used the example.conf that was inside the folder but it's all way over my head, could anybody create a code i can copy/past or atleast get me started so i have a better hands-on understanding?
A shortcut won't work unless you're "in" the directory where the kernel files are kept. If you're writing a .bat file make sure to cd into the cgminer directory before starting and make SURE all the files it came with are there. Also make sure to use the latest version (currently 2.2.6).
|
|
|
export DISPLAY=:0
Also you will get more overclocking ability with 11.6 instead of 11.4.
|
|
|
cgminer 2.2.6 out. 2+ more mhash on 7970. It's getting harder and harder to extract much more 1200/1050+5% clocks, intensity 11 - 717 Mhash.
|
|
|
Just installed 11.6 on 64-bit windows 7, and cgminer 2.2.3 was only showing 1 GPU. I updated to cgminer 2.2.6, and now I get this: Too many values passed to set temp cutoff Do I need dummy plugs for the cards, or is there some other issue? Should I just suck it up and install the latest drivers? No monitors or dummy plugs are connected. I also tried the SDK 2.4 standalone installer, and get the same results. 11.6 needs dummy plugs on windows. You need a later driver but beware of SDK fuckage.
|
|
|
New release: Version 2.2.6 - February 16, 2012
Links in top post as always.
Trying hard to make the most current release as stable as possible at the moment.
Updated customised poclbm kernel for 7970 worth 2 more Mhash, and same on any other GPUs when run on SDK2.6. BIG FAT WARNINGS if SDK 2.6 is detected. Ability to tune dynamic mode with a new --gpu-dyninterval parameter. This tells cgminer how long maximum it can keep the GPU busy for. Default was 7ms but can now be set to any value. Setting it higher will increase your hashrate and make your desktop less responsive. This is in response to requests to improve hashrate back to what it was when 2 threads were in use. The display should keep all logs now from startup when the curses formatted display begins. The bug where dynamic mode increases stale rate by submitting duplicate work should be fixed.
Full changelog:
Provide warning on each startup about sdk 2.6 - Fix unused warnings on win32. - bitforce: Simplify BFopen WIN32 ifdef/else - Fix initialization warning with jansson 1.3 - bitforce: Cleanup extraneous TODO that isn't needed - Move tcsetattr (and new tcflush) into *nix BFopen to simplify things a bit - Add message explaining 2nd thread disabling for dynamic mode and how to tune it. - Move logwindow down once number of devices is known. - Automatically choose phatk kernel for bitalign non-gcn ATI cards, and then only select poclbm if SDK2.6 is detected. - Allow the refresh interval to be adjusted in dynamic intensity with a --gpu-dyninterval parameter. - Make curses display visible right from the beginning and fix the window sizes so the initial messages don't get lost once the status window is drawn. - The amount of work scanned can fluctuate when intensity changes and since we do this one cycle behind, we increment the work more than enough to prevent repeati - bitforce: Set a 30 second timeout for serial port on Windows, since the default is undefined - Use PreVal4addT1 instead of PreVal4 in poclbm kernel. - Import PreVal4 and PreVal0 into poclbm kernel. - Import more prepared constants into poclbm kernel. - Keep variables in one array but use Vals[] name for consistency with other kernel designs. - Replace constants that are mandatorily added in poclbm kernel with one value. - Remove addition of final constant before testing for result in poclbm kernel. - Hand optimise variable addition order. - Hand optimise first variable declaration order in poclbm kernel. - Radical reordering machine based first pass to change variables as late as possible, bringing their usage close together. - fix strcpy NULL pointer if env HOME unset. - bitforce: Disable automatic scanning when at least one device is specified manually - Unroll all poclbm additions to enable further optimisations.
|
|
|
Hi CK, have you found the problem causing the high % of rejected shares? i set it to -I 9 and still having way to many rejects, it starts very ugly but with time appears to get better, my current run shows (A:35107 R:2059), less than with dynamic set but still. I'm running 2.2.4 i didn't notice that in 2.0.8
This is the first time you've reported it? Suggestions: Perhaps your pool is actually asking you to submit the old shares even if they're stale. Newer cgminer supports the "submitold" feature now. I get .3-.4% rejects with static intensities so I can't reproduce your problem here. p2pool also has a significantly higher reject rate if that's what you're on/have moved to.
|
|
|
Hmm if I didn't know any better, it looks like the auto-fan is not being enabled on the other devices. Can you click on GPU setting, Change setting and select each device in order and see if they report: Fan autotune is enabled (0-85) Thanks "Fan autotune is enabled (0-85)" is reported for each of the 6 cores. Thanks. In the meantime, I've audited the code and can't find where this might happen, so a fix for this is not coming in the next version I'm afraid. I'll keep looking.
|
|
|
3x5970 -- why is my GPU 4/5 fan speed so high?
This is after an hour or so of 2.2.5, but the fan speeds were the same after 70 hours of 2.2.4.
While it's running, can you go into settings and enable debug mode and see what it's saying about temperatures and fan speeds? It will spew all sorts of info so you'll have to watch for a while to find it. Sequences like this were output at 3 or 6 second intervals, so there were dozens during the sample period. 100% of them had fanspeed changes for GPU 2, with temperatures rising or falling and setting its speed to 68 or 67. No other GPU was mentioned in a "fanspeed" message. 2.2.5+ (commit 158c09107cb9ec1657c42213e3d7e5c972ff0aae; Tue Feb 14 13:33:17 2012 UTC) Hmm if I didn't know any better, it looks like the auto-fan is not being enabled on the other devices. Can you click on GPU setting, Change setting and select each device in order and see if they report: Fan autotune is enabled (0-85) Thanks
|
|
|
3x5970 -- why is my GPU 4/5 fan speed so high?
This is after an hour or so of 2.2.5, but the fan speeds were the same after 70 hours of 2.2.4.
While it's running, can you go into settings and enable debug mode and see what it's saying about temperatures and fan speeds? It will spew all sorts of info so you'll have to watch for a while to find it.
|
|
|
Is this normal for P2Pool if running on a different machine on the same P2P node? You have hardware errors with cgminer which means something is horribly wrong with your install and this has nothing to do with p2pool. Start with a fresh install of cgminer (2.2.5 is latest).
|
|
|
|