You want to crawl 524620 pages? I'm afraid it's impossible if you're using customer desktop because :
1. Each page have size about 100KB (only HTML code tag, excluding all external files) or about 50GB in total
2. If they have protection against DDoS, bot or spammer, you'd be banned before you could crawl 500K pages
IMO contacting them to ask JSON/dumped SQL data is more realistic option.But it's one good website in this case.
it is possible
1- with idm go to tasks > add batch download>from 0 to 1000> ok (about 110MB) (50*1000=50k top bitcoin addresses)
concatenating all the htmls files in one html edit with notepad++ the final file and make sure to only keep the addresses in first column
(you need to repeat this many times)
i did try this ... bored and stop