Bitcoin Forum
June 21, 2024, 11:08:54 PM *
News: Voting for pizza day contest
 
   Home   Help Search Login Register More  
Pages: « 1 ... 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 [392] 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 ... 1315 »
  Print  
Author Topic: [ANN][BURST] Burst | Efficient HDD Mining | New 1.2.3 Fork block 92000  (Read 2170605 times)
SpeedDemon13
Hero Member
*****
Offline Offline

Activity: 518
Merit: 500



View Profile WWW
September 09, 2014, 08:17:38 AM
 #7821

anybody know how long can a CPU last running at 95-100% 24/7  ?

Can run fine for a long period, depending on the cooling and if the cpu is overclocked too high/too long.

stock , no extra cooling , can run for at least a year ?

Don't know if a year, but maybe for a couple of months it could. Server cpus are probably more tolerable to 95%~100% loads than the common desktop cpus to run for a year.

Regardless, it should run with no issues if everything around it runs properly. ie. motherboard, psu, etc....


oh people, that makes me cringe.

My Mail server ist an ancient 23 year old box, still humming away with up-to-date software. Load is at 100%, because it is still running the dnetc-client from 2003 onwards. Not that that make any particular sense.

Anyway, if your machine has no massive design flaws (power supply at >90%, inadequate cooling, bad manufacturing (heat transfer, bad capacitors, oxidizing contacts) it should keep on running for years, even at 100% load.

Stop irritating the rookies, please.


You didn't account for one variable that kills most hardware, dust. A lot of people don't do that maintenance cleaning and that creates heat, possible failure.

I'm not trying to irritate anyone, if that what your assumption of what I'm doing.

CRYPTSY exchange: https://www.cryptsy.com/users/register?refid=9017 BURST= BURST-TE3W-CFGH-7343-6VM6R BTC=1CNsqGUR9YJNrhydQZnUPbaDv6h4uaYCHv ETH=0x144bc9fe471d3c71d8e09d58060d78661b1d4f32 SHF=0x13a0a2cb0d55eca975cf2d97015f7d580ce52d85 EXP=0xd71921dca837e415a58ca0d6dd2223cc84e0ea2f SC=6bdf9d12a983fed6723abad91a39be4f95d227f9bdb0490de3b8e5d45357f63d564638b1bd71 CLAMS=xGVTdM9EJpNBCYAjHFVxuZGcqvoL22nP6f SOIL=0x8b5c989bc931c0769a50ecaf9ffe490c67cb5911
coinits
Legendary
*
Offline Offline

Activity: 1582
Merit: 1019


011110000110110101110010


View Profile
September 09, 2014, 08:23:30 AM
 #7822

anybody know how long can a CPU last running at 95-100% 24/7  ?

Can run fine for a long period, depending on the cooling and if the cpu is overclocked too high/too long.

stock , no extra cooling , can run for at least a year ?

Don't know if a year, but maybe for a couple of months it could. Server cpus are probably more tolerable to 95%~100% loads than the common desktop cpus to run for a year.

Regardless, it should run with no issues if everything around it runs properly. ie. motherboard, psu, etc....


oh people, that makes me cringe.

My Mail server ist an ancient 23 year old box, still humming away with up-to-date software. Load is at 100%, because it is still running the dnetc-client from 2003 onwards. Not that that make any particular sense.

Anyway, if your machine has no massive design flaws (power supply at >90%, inadequate cooling, bad manufacturing (heat transfer, bad capacitors, oxidizing contacts) it should keep on running for years, even at 100% load.

Stop irritating the rookies, please.


You didn't account for one variable that kills most hardware, dust. A lot of people don't do that maintenance cleaning and that creates heat, possible failure.

I'm not trying to irritate anyone, if that what your assumption of what I'm doing.

Dust is a serial computer killer!

Jump you fuckers! | The thing about smart motherfuckers is they sound like crazy motherfuckers to dumb motherfuckers. | My sig space for rent for 0.01 btc per week.
go6ooo1212
Legendary
*
Offline Offline

Activity: 1512
Merit: 1000


quarkchain.io


View Profile
September 09, 2014, 08:25:55 AM
 #7823

Im running it with Win8.1 Pro-64bit and I cant exseed the memory usage above 2.6GB. When I try to change the value ot run_generate.bat above 1000 , it gives me an error. Does someone have an Idea , this is my comand line:
C:\Windows\SysWOW64\java -Xmx1000m -cp pocminer.jar;lib/*;lib/akka/*;lib/jetty/* pocminer.POCMiner generate "address" 4096 7168000 2048 5
bipben
Member
**
Offline Offline

Activity: 60
Merit: 10


View Profile
September 09, 2014, 08:26:22 AM
 #7824

Can't get it work and What is reco <threads> for R9 Series Card ?

Hi everyone,

After many hours of setup I finally made it. I have a 1Tb generation in progress and 3x100Gb already finished.
I would like to test the V2 pool but I haven't any BURST for now. Could someone send me 1 BURST to test it please ? Here is my address : BURST-YA29-QCEW-QXC3-BKXDL.

Regarding the plot generation, I found an OpenCL implementation of Shabal (https://github.com/aznboy84/X15GPU/blob/master/kernel/shabal.cl) that could be used to make a GPU version of the generator. I will try to work on it when I have some free time.

Regards

Hi everyone,

As promised I have been working on a GPU plot generator on the last few days. I made a little program built on top of OpenCL, and it seems to work pretty well in CPU mode. Unfortunately, I can't test the GPU mode as it requires a very powerfull graphic card (with at least 46kB private memory per compute unit, because the algorithm needs at least 4096*64 static bytes to store an entire plot).

Here is a preview you can test for now :
gpuPlotGenerator-src-1.0.0.7z : https://mega.co.nz/#!bcF2yKKL!3Ud86GaibgvwBehoxkbO4UNdiBgsaixRx7ksHrgNbDI
gpuPlotGenerator-bin-win-x86-1.0.0.7z : https://mega.co.nz/#!HJsziTCK!UmAMoEHQ3z34R4RsXoIkYo9rYd4LnFtO_pw-R4KObJs

I will build another release in the end of the day with some minor improvements (threads per compute unit selection, output of OpenCL error codes, improvement of the Makefile to generate the distribution directly).
I will also try to figure out another mean to dispatch the work between the GPU threads to reduce the amount of private memory needed by the program.

For the windows people, you can use the binary version directly.
For the linux people, just download the source archive, make sure to modify the OpenCL library and lib path in the makefile (and maybe the executable name), and build the project via "make". To run the program, you need the "kernel" and the "plots" directories beside the executable.

The executable usage is : ./gpuPlotGenerator <address> <start nonce> <nonces> <stagger size>
The parameters are the same as the original plot generator, without the threads number.

If you find bugs or if you want some new features, let me now.

If you want to support me, here are my Bitcoin and Burst addresses :
Bitcoin: 138gMBhCrNkbaiTCmUhP9HLU9xwn5QKZgD
Burst: BURST-YA29-QCEW-QXC3-BKXDL

Regards

Unfortunately, I can't test the GPU mode as it requires a very powerfull graphic card (with at least 46kB private memory per compute unit, because the algorithm needs at least 4096*64 static bytes to store an entire plot).
It's nice to see someone else working on this, since I seem to have failed in it.

Private memory is actually part of global on AMD cards, so storing it in private isn't any better than just using global for everything; it's local that needs to aimed for for the massive speedup. No AMD cards have more than 64KB local per workgroup, which makes storing it all in local impossible however.

I haven't tried your implementation yet, but on my own first attempt, I also used global on everything also, and the result was faster than the java plotter, but slower than dcct's c plotter. My 2nd attempt used a 32KB local buffer I rotated through for storing the currently being hashed stuff, however I couldn't figure out how to get it copied also to global fast enough, and the local -> global copy killed the performance.

You might be interested in those kernels here: https://bitcointalk.org/index.php?topic=731923.msg8695829#msg8695829

Thanks, I will look at your kernels to see if I can find a better solution.

Here is the new version. I reduced the amount of memory used from 40KB to about 1KB per unit. The only drawback is that it requires twice the global memory as before. I will search a mean to reduce this overhead later.
In CPU mode, it all goes pretty well (when no graphic card is detected).
The GPU mode is still kind of buggy on my graphic card (an old GeForce 9300M GS), don't know the exact reason yet. Sometimes it works, sometimes not. I will try to fix this issue tomorrow.

Here are the files :
gpuPlotGenerator-src-1.1.0.7z : https://mega.co.nz/#!iYFWAL5B!BvtmRQ5qGq4gGwjDglFNtDtNIX4LDaUvATBtClBdTlQ
gpuPlotGenerator-bin-win-x86-1.1.0.7z : https://mega.co.nz/#!aBVGBBQD!tBsRtb8VrHR12_anrFTrl41U0fPQu_OqFnxyi5nCyBY

For the linux users, the Makefile has a new target named "dist" that builds and copy/paste all the necessary files to the "bin" directory.

The executable usage is : ./gpuPlotGenerator <path> <address> <start nonce> <nonces> <stagger size> <threads>
<path> : the path to the plots directory
<threads> : number of parrallel threads for each work group

Hi,

I am aware of the current bug and I will work on it today.
For the <threads> parameter, use a multiple of 64, the typical value is 256 for most of the graphic cards.

So, the <threads> parameter is not needed for Windows exe? Is the <address> parameter the numerical address created from the passphrase? How does this gpu plotter perform vs the cpu plotter? Are the plots error free or is there a margin of error compared to the cpu plotted ones?

Update: Didn't read it properly about the thread parameter, understand it's required for Windows and Linux. Why is the multiples 64 (same as the scoop size) and typical value 256 (same as the plot size) for gpu's?

For the <address> parameter, yes it's the numerical one.
Once finished, the plotter will generate the exact same file as the CPU plotter, but faster I hope. There is no performance comparison available for now as I am still correcting it.
For the <threads> parameter, it depends on your graphic card architecture. Most of the graphic cards perform computation with workgroups of 64*N threads (256 most of the time). The maximum workgroup size is available on the manufacturer website for each model.

Burst: BURST-YA29-QCEW-QXC3-BKXDL
alphateam
Hero Member
*****
Offline Offline

Activity: 1036
Merit: 531


View Profile
September 09, 2014, 08:27:27 AM
 #7825

What is your full parameter ?

Hi everyone,

After many hours of setup I finally made it. I have a 1Tb generation in progress and 3x100Gb already finished.
I would like to test the V2 pool but I haven't any BURST for now. Could someone send me 1 BURST to test it please ? Here is my address : BURST-YA29-QCEW-QXC3-BKXDL.

Regarding the plot generation, I found an OpenCL implementation of Shabal (https://github.com/aznboy84/X15GPU/blob/master/kernel/shabal.cl) that could be used to make a GPU version of the generator. I will try to work on it when I have some free time.

Regards

Hi everyone,

As promised I have been working on a GPU plot generator on the last few days. I made a little program built on top of OpenCL, and it seems to work pretty well in CPU mode. Unfortunately, I can't test the GPU mode as it requires a very powerfull graphic card (with at least 46kB private memory per compute unit, because the algorithm needs at least 4096*64 static bytes to store an entire plot).

Here is a preview you can test for now :
gpuPlotGenerator-src-1.0.0.7z : https://mega.co.nz/#!bcF2yKKL!3Ud86GaibgvwBehoxkbO4UNdiBgsaixRx7ksHrgNbDI
gpuPlotGenerator-bin-win-x86-1.0.0.7z : https://mega.co.nz/#!HJsziTCK!UmAMoEHQ3z34R4RsXoIkYo9rYd4LnFtO_pw-R4KObJs

I will build another release in the end of the day with some minor improvements (threads per compute unit selection, output of OpenCL error codes, improvement of the Makefile to generate the distribution directly).
I will also try to figure out another mean to dispatch the work between the GPU threads to reduce the amount of private memory needed by the program.

For the windows people, you can use the binary version directly.
For the linux people, just download the source archive, make sure to modify the OpenCL library and lib path in the makefile (and maybe the executable name), and build the project via "make". To run the program, you need the "kernel" and the "plots" directories beside the executable.

The executable usage is : ./gpuPlotGenerator <address> <start nonce> <nonces> <stagger size>
The parameters are the same as the original plot generator, without the threads number.

If you find bugs or if you want some new features, let me now.

If you want to support me, here are my Bitcoin and Burst addresses :
Bitcoin: 138gMBhCrNkbaiTCmUhP9HLU9xwn5QKZgD
Burst: BURST-YA29-QCEW-QXC3-BKXDL

Regards

Unfortunately, I can't test the GPU mode as it requires a very powerfull graphic card (with at least 46kB private memory per compute unit, because the algorithm needs at least 4096*64 static bytes to store an entire plot).
It's nice to see someone else working on this, since I seem to have failed in it.

Private memory is actually part of global on AMD cards, so storing it in private isn't any better than just using global for everything; it's local that needs to aimed for for the massive speedup. No AMD cards have more than 64KB local per workgroup, which makes storing it all in local impossible however.

I haven't tried your implementation yet, but on my own first attempt, I also used global on everything also, and the result was faster than the java plotter, but slower than dcct's c plotter. My 2nd attempt used a 32KB local buffer I rotated through for storing the currently being hashed stuff, however I couldn't figure out how to get it copied also to global fast enough, and the local -> global copy killed the performance.

You might be interested in those kernels here: https://bitcointalk.org/index.php?topic=731923.msg8695829#msg8695829

Thanks, I will look at your kernels to see if I can find a better solution.

Here is the new version. I reduced the amount of memory used from 40KB to about 1KB per unit. The only drawback is that it requires twice the global memory as before. I will search a mean to reduce this overhead later.
In CPU mode, it all goes pretty well (when no graphic card is detected).
The GPU mode is still kind of buggy on my graphic card (an old GeForce 9300M GS), don't know the exact reason yet. Sometimes it works, sometimes not. I will try to fix this issue tomorrow.

Here are the files :
gpuPlotGenerator-src-1.1.0.7z : https://mega.co.nz/#!iYFWAL5B!BvtmRQ5qGq4gGwjDglFNtDtNIX4LDaUvATBtClBdTlQ
gpuPlotGenerator-bin-win-x86-1.1.0.7z : https://mega.co.nz/#!aBVGBBQD!tBsRtb8VrHR12_anrFTrl41U0fPQu_OqFnxyi5nCyBY

For the linux users, the Makefile has a new target named "dist" that builds and copy/paste all the necessary files to the "bin" directory.

The executable usage is : ./gpuPlotGenerator <path> <address> <start nonce> <nonces> <stagger size> <threads>
<path> : the path to the plots directory
<threads> : number of parrallel threads for each work group

So the usage would be like this: "D:/gpuPlotGenerator <numerical_account_address> 0  819200 4096 <cpu/gpu_threads?>"

Is that format correct? Is the thread count need for gpu plotting(Point out in bold)? What's the nonce/minute rate?

Hi,

This is still a buggy early stage version. I post it here to have feedback from people who owns more powerfull graphic cards (the behaviour may vary from one card to another).
But yes, the final usage would be the one you mentioned. The threads parameter is the number of threads used in the local work group. In GPU mode, the value should be a multiple a 64, 256 is the typical value for most of the cards.


Ok i made a test with my R9 290

I Put 256 in thread (apparently can't put more)

And in 1min15 i generate from nonce 888597 to nonce 900885, So 9830 nonce minute, not bad at all


This is my parameter

Code:
gpuPlotGenerator.exe plots myaccount 68847637 18500000 4096 256
alphateam
Hero Member
*****
Offline Offline

Activity: 1036
Merit: 531


View Profile
September 09, 2014, 08:29:27 AM
 #7826



What's your gpu load and temps?

Can't say, when i launch monitoring drivers crash. But card seems to not make noise and not hot
Finley
Full Member
***
Offline Offline

Activity: 136
Merit: 100


View Profile
September 09, 2014, 08:29:32 AM
 #7827

Hi dev have you ever ccosidered to join the supernet work?
alphateam
Hero Member
*****
Offline Offline

Activity: 1036
Merit: 531


View Profile
September 09, 2014, 08:30:43 AM
 #7828



Hi

How exactly do you plot in 'GPU mode'?

I also have an R9 290 but was told that GPU plotting wasn't a thing.

Just made a batch file in the same folder with that :

Code:
gpuPlotGenerator.exe plots myaccount 68847637 18500000 4096 256
coinits
Legendary
*
Offline Offline

Activity: 1582
Merit: 1019


011110000110110101110010


View Profile
September 09, 2014, 08:31:22 AM
 #7829

http://burst.cryptoport.io/acc/2715798095717378439 - Now there is some mining power!

Jump you fuckers! | The thing about smart motherfuckers is they sound like crazy motherfuckers to dumb motherfuckers. | My sig space for rent for 0.01 btc per week.
LivsUA
Full Member
***
Offline Offline

Activity: 210
Merit: 100



View Profile
September 09, 2014, 08:31:40 AM
 #7830

What is your full parameter ?

Hi everyone,

After many hours of setup I finally made it. I have a 1Tb generation in progress and 3x100Gb already finished.
I would like to test the V2 pool but I haven't any BURST for now. Could someone send me 1 BURST to test it please ? Here is my address : BURST-YA29-QCEW-QXC3-BKXDL.

Regarding the plot generation, I found an OpenCL implementation of Shabal (https://github.com/aznboy84/X15GPU/blob/master/kernel/shabal.cl) that could be used to make a GPU version of the generator. I will try to work on it when I have some free time.

Regards

Hi everyone,

As promised I have been working on a GPU plot generator on the last few days. I made a little program built on top of OpenCL, and it seems to work pretty well in CPU mode. Unfortunately, I can't test the GPU mode as it requires a very powerfull graphic card (with at least 46kB private memory per compute unit, because the algorithm needs at least 4096*64 static bytes to store an entire plot).

Here is a preview you can test for now :
gpuPlotGenerator-src-1.0.0.7z : https://mega.co.nz/#!bcF2yKKL!3Ud86GaibgvwBehoxkbO4UNdiBgsaixRx7ksHrgNbDI
gpuPlotGenerator-bin-win-x86-1.0.0.7z : https://mega.co.nz/#!HJsziTCK!UmAMoEHQ3z34R4RsXoIkYo9rYd4LnFtO_pw-R4KObJs

I will build another release in the end of the day with some minor improvements (threads per compute unit selection, output of OpenCL error codes, improvement of the Makefile to generate the distribution directly).
I will also try to figure out another mean to dispatch the work between the GPU threads to reduce the amount of private memory needed by the program.

For the windows people, you can use the binary version directly.
For the linux people, just download the source archive, make sure to modify the OpenCL library and lib path in the makefile (and maybe the executable name), and build the project via "make". To run the program, you need the "kernel" and the "plots" directories beside the executable.

The executable usage is : ./gpuPlotGenerator <address> <start nonce> <nonces> <stagger size>
The parameters are the same as the original plot generator, without the threads number.

If you find bugs or if you want some new features, let me now.

If you want to support me, here are my Bitcoin and Burst addresses :
Bitcoin: 138gMBhCrNkbaiTCmUhP9HLU9xwn5QKZgD
Burst: BURST-YA29-QCEW-QXC3-BKXDL

Regards

Unfortunately, I can't test the GPU mode as it requires a very powerfull graphic card (with at least 46kB private memory per compute unit, because the algorithm needs at least 4096*64 static bytes to store an entire plot).
It's nice to see someone else working on this, since I seem to have failed in it.

Private memory is actually part of global on AMD cards, so storing it in private isn't any better than just using global for everything; it's local that needs to aimed for for the massive speedup. No AMD cards have more than 64KB local per workgroup, which makes storing it all in local impossible however.

I haven't tried your implementation yet, but on my own first attempt, I also used global on everything also, and the result was faster than the java plotter, but slower than dcct's c plotter. My 2nd attempt used a 32KB local buffer I rotated through for storing the currently being hashed stuff, however I couldn't figure out how to get it copied also to global fast enough, and the local -> global copy killed the performance.

You might be interested in those kernels here: https://bitcointalk.org/index.php?topic=731923.msg8695829#msg8695829

Thanks, I will look at your kernels to see if I can find a better solution.

Here is the new version. I reduced the amount of memory used from 40KB to about 1KB per unit. The only drawback is that it requires twice the global memory as before. I will search a mean to reduce this overhead later.
In CPU mode, it all goes pretty well (when no graphic card is detected).
The GPU mode is still kind of buggy on my graphic card (an old GeForce 9300M GS), don't know the exact reason yet. Sometimes it works, sometimes not. I will try to fix this issue tomorrow.

Here are the files :
gpuPlotGenerator-src-1.1.0.7z : https://mega.co.nz/#!iYFWAL5B!BvtmRQ5qGq4gGwjDglFNtDtNIX4LDaUvATBtClBdTlQ
gpuPlotGenerator-bin-win-x86-1.1.0.7z : https://mega.co.nz/#!aBVGBBQD!tBsRtb8VrHR12_anrFTrl41U0fPQu_OqFnxyi5nCyBY

For the linux users, the Makefile has a new target named "dist" that builds and copy/paste all the necessary files to the "bin" directory.

The executable usage is : ./gpuPlotGenerator <path> <address> <start nonce> <nonces> <stagger size> <threads>
<path> : the path to the plots directory
<threads> : number of parrallel threads for each work group

So the usage would be like this: "D:/gpuPlotGenerator <numerical_account_address> 0  819200 4096 <cpu/gpu_threads?>"

Is that format correct? Is the thread count need for gpu plotting(Point out in bold)? What's the nonce/minute rate?

Hi,

This is still a buggy early stage version. I post it here to have feedback from people who owns more powerfull graphic cards (the behaviour may vary from one card to another).
But yes, the final usage would be the one you mentioned. The threads parameter is the number of threads used in the local work group. In GPU mode, the value should be a multiple a 64, 256 is the typical value for most of the cards.


Ok i made a test with my R9 290

I Put 256 in thread (apparently can't put more)

And in 1min15 i generate from nonce 888597 to nonce 900885, So 9830 nonce minute, not bad at all

What config you use???
I have tried this on my 280x:

--->  gpuPlotGenerator.exe "C:\Users\Fantom\Desktop\GPU_POC\plots" NUMERICALADRESS 0 16382 4096 4

and it just exits without writing new file..

bipben
Member
**
Offline Offline

Activity: 60
Merit: 10


View Profile
September 09, 2014, 08:32:40 AM
 #7831

What is your full parameter ?

Hi everyone,

After many hours of setup I finally made it. I have a 1Tb generation in progress and 3x100Gb already finished.
I would like to test the V2 pool but I haven't any BURST for now. Could someone send me 1 BURST to test it please ? Here is my address : BURST-YA29-QCEW-QXC3-BKXDL.

Regarding the plot generation, I found an OpenCL implementation of Shabal (https://github.com/aznboy84/X15GPU/blob/master/kernel/shabal.cl) that could be used to make a GPU version of the generator. I will try to work on it when I have some free time.

Regards

Hi everyone,

As promised I have been working on a GPU plot generator on the last few days. I made a little program built on top of OpenCL, and it seems to work pretty well in CPU mode. Unfortunately, I can't test the GPU mode as it requires a very powerfull graphic card (with at least 46kB private memory per compute unit, because the algorithm needs at least 4096*64 static bytes to store an entire plot).

Here is a preview you can test for now :
gpuPlotGenerator-src-1.0.0.7z : https://mega.co.nz/#!bcF2yKKL!3Ud86GaibgvwBehoxkbO4UNdiBgsaixRx7ksHrgNbDI
gpuPlotGenerator-bin-win-x86-1.0.0.7z : https://mega.co.nz/#!HJsziTCK!UmAMoEHQ3z34R4RsXoIkYo9rYd4LnFtO_pw-R4KObJs

I will build another release in the end of the day with some minor improvements (threads per compute unit selection, output of OpenCL error codes, improvement of the Makefile to generate the distribution directly).
I will also try to figure out another mean to dispatch the work between the GPU threads to reduce the amount of private memory needed by the program.

For the windows people, you can use the binary version directly.
For the linux people, just download the source archive, make sure to modify the OpenCL library and lib path in the makefile (and maybe the executable name), and build the project via "make". To run the program, you need the "kernel" and the "plots" directories beside the executable.

The executable usage is : ./gpuPlotGenerator <address> <start nonce> <nonces> <stagger size>
The parameters are the same as the original plot generator, without the threads number.

If you find bugs or if you want some new features, let me now.

If you want to support me, here are my Bitcoin and Burst addresses :
Bitcoin: 138gMBhCrNkbaiTCmUhP9HLU9xwn5QKZgD
Burst: BURST-YA29-QCEW-QXC3-BKXDL

Regards

Unfortunately, I can't test the GPU mode as it requires a very powerfull graphic card (with at least 46kB private memory per compute unit, because the algorithm needs at least 4096*64 static bytes to store an entire plot).
It's nice to see someone else working on this, since I seem to have failed in it.

Private memory is actually part of global on AMD cards, so storing it in private isn't any better than just using global for everything; it's local that needs to aimed for for the massive speedup. No AMD cards have more than 64KB local per workgroup, which makes storing it all in local impossible however.

I haven't tried your implementation yet, but on my own first attempt, I also used global on everything also, and the result was faster than the java plotter, but slower than dcct's c plotter. My 2nd attempt used a 32KB local buffer I rotated through for storing the currently being hashed stuff, however I couldn't figure out how to get it copied also to global fast enough, and the local -> global copy killed the performance.

You might be interested in those kernels here: https://bitcointalk.org/index.php?topic=731923.msg8695829#msg8695829

Thanks, I will look at your kernels to see if I can find a better solution.

Here is the new version. I reduced the amount of memory used from 40KB to about 1KB per unit. The only drawback is that it requires twice the global memory as before. I will search a mean to reduce this overhead later.
In CPU mode, it all goes pretty well (when no graphic card is detected).
The GPU mode is still kind of buggy on my graphic card (an old GeForce 9300M GS), don't know the exact reason yet. Sometimes it works, sometimes not. I will try to fix this issue tomorrow.

Here are the files :
gpuPlotGenerator-src-1.1.0.7z : https://mega.co.nz/#!iYFWAL5B!BvtmRQ5qGq4gGwjDglFNtDtNIX4LDaUvATBtClBdTlQ
gpuPlotGenerator-bin-win-x86-1.1.0.7z : https://mega.co.nz/#!aBVGBBQD!tBsRtb8VrHR12_anrFTrl41U0fPQu_OqFnxyi5nCyBY

For the linux users, the Makefile has a new target named "dist" that builds and copy/paste all the necessary files to the "bin" directory.

The executable usage is : ./gpuPlotGenerator <path> <address> <start nonce> <nonces> <stagger size> <threads>
<path> : the path to the plots directory
<threads> : number of parrallel threads for each work group

So the usage would be like this: "D:/gpuPlotGenerator <numerical_account_address> 0  819200 4096 <cpu/gpu_threads?>"

Is that format correct? Is the thread count need for gpu plotting(Point out in bold)? What's the nonce/minute rate?

Hi,

This is still a buggy early stage version. I post it here to have feedback from people who owns more powerfull graphic cards (the behaviour may vary from one card to another).
But yes, the final usage would be the one you mentioned. The threads parameter is the number of threads used in the local work group. In GPU mode, the value should be a multiple a 64, 256 is the typical value for most of the cards.


Ok i made a test with my R9 290

I Put 256 in thread (apparently can't put more)

And in 1min15 i generate from nonce 888597 to nonce 900885, So 9830 nonce minute, not bad at all

Wow! So it really works on some models after all! Glad to read it. I am still investigating to correct the bug that occurs on the other graphic cards.
Thank you for your feedback.

Burst: BURST-YA29-QCEW-QXC3-BKXDL
AizenSou
Hero Member
*****
Offline Offline

Activity: 938
Merit: 1000


View Profile
September 09, 2014, 08:33:04 AM
 #7832

While I'm getting ready for Storj when beta is out, I decided to play with burst a bit.
I've read a lot of this thread and most things are pretty clear, but I have 2 questions remaining.
How do you tell the java miner to consider plots present on a different drive letter (windows)?  can someone point me to the syntax for this?
When I downloaded the plot.exe for windows to a sandbox using chrome, it complained of malware and suspended the download, why?
thanks

- You need to copy the java miner to the disk you want to generate plots, the miner always generate plots on plots folder of current path. Pretty easy fix to change this, but I don't think it's a priority.
- False positive is pretty standard today with miner stuffs, don't you think? Wink
I was wondering about mining, not plot generation.  is it possible to run one miner instance but have it read plotted files from multiple drives?
I typically don't trust anything that says click here to download regardless of false positives.  How come the Linux plotter is linked on OP but windows is not?

Sorry I misread your question. But in general these 2 tools work in the same manner.
- The java miner won't work with multiple folders. In Linux you could use symlink and it works for me, I dunno Windows will work with linking ? I don't have Windows so I can't test. In worst case you just copy the miner to your other places and start it from there.
- The plot from dcct is originally written for Linux. Someone provides a binary for Windows so of course you should be cautious. Sorry but I can't help you with this question.
alphateam
Hero Member
*****
Offline Offline

Activity: 1036
Merit: 531


View Profile
September 09, 2014, 08:34:11 AM
 #7833

http://burst.cryptoport.io/acc/2715798095717378439 - Now there is some mining power!

Wow how many TB?
alphateam
Hero Member
*****
Offline Offline

Activity: 1036
Merit: 531


View Profile
September 09, 2014, 08:35:05 AM
 #7834



What config you use???
I have tried this on my 280x:

--->  gpuPlotGenerator.exe "C:\Users\Fantom\Desktop\GPU_POC\plots" NUMERICALADRESS 0 16382 4096 4

and it just exits without writing new file..

Code:
gpuPlotGenerator.exe plots myaccount 68847637 18500000 4096 256
fanepatent
Full Member
***
Offline Offline

Activity: 224
Merit: 100


View Profile
September 09, 2014, 08:35:51 AM
 #7835

http://www.aliexpress.com/item/PCI-e-Internal-8-Port-SATA-iii-3-0-6Gb-PCI-Express-SATA3-0-Controller-Card/1935753227.html

BURST - BURST-58XP-63WY-XSVQ-ASG9A
alphateam
Hero Member
*****
Offline Offline

Activity: 1036
Merit: 531


View Profile
September 09, 2014, 08:37:07 AM
 #7836

Proof that gpu plotter works for me

newuser01
Sr. Member
****
Offline Offline

Activity: 434
Merit: 250


View Profile
September 09, 2014, 08:39:50 AM
 #7837

Proof that gpu plotter works for me



290x?

I cant get it working on my 280x card
fabula
Hero Member
*****
Offline Offline

Activity: 644
Merit: 500



View Profile
September 09, 2014, 08:40:59 AM
 #7838

300 satoshi today. Not stable at all.
10 days and price still not rise of 1 satoshi.
alphateam
Hero Member
*****
Offline Offline

Activity: 1036
Merit: 531


View Profile
September 09, 2014, 08:44:36 AM
 #7839

Proof that gpu plotter works for me



290x?

I cant get it working on my 280x card

Yep right, i post my config some post before
burstcoin (OP)
Sr. Member
****
Offline Offline

Activity: 280
Merit: 250


View Profile
September 09, 2014, 08:45:13 AM
 #7840

http://burst.cryptoport.io/acc/2715798095717378439 - Now there is some mining power!

Looks like he's 15.5% of the hashpower on my v2 pool.

BURST-QHCJ-9HB5-PTGC-5Q8J9
Pages: « 1 ... 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 [392] 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 ... 1315 »
  Print  
 
Jump to:  

Powered by MySQL Powered by PHP Powered by SMF 1.1.19 | SMF © 2006-2009, Simple Machines Valid XHTML 1.0! Valid CSS!