Hi,
Sorry, I posted this in the Einstein threads by mistake. Can the mods please delete that ??
I have no Einstein tasks running or waiting to upload, but, the disk usage is up at 90 MB. What's the go ?? Can I clear that disk usage ??? How do I find where that disk usage is, so I can clear it ??? Is this normal ???
Cheers,
David
Copyright © 2024 Einstein@Home. All rights reserved.
Excessive Disk Usage
)
If all your Einstein tasks are uploaded and reported, you could reset the project. That would delete all Einstein@home-related files.
However, if you want to resume working for Einstein, all executables and (semi)static data files will have to be downloaded again.
Gruß,
Gundolf
Computer sind nicht alles im Leben. (Kleiner Scherz)
Thanks Gundolf, I'll
)
Thanks Gundolf,
I'll download tomorrow morning when I'm on off-peak.
Cheers,
David
RE: Thanks Gundolf, I'll
)
Um.... Does that also clear out my credits ??? Or do they remain intact ??
RE: Um.... Does that
)
Credits remain.
Michael
Team Linux Users Everywhere
Cool... Thanks. Zap.
)
Cool... Thanks.
Zap.
I think that 90MB isn't that
)
I think that 90MB isn't that much for a S5R6 task. Locality scheduling should make sure that you'll get new tasks that will use files that already are on your machine as much as possible, to minimize your new download volume. If you delete the files now, you'll probably download a very similar set of files again with your next S5R6 task.
BM
BM
It's now up to 237MB after
)
It's now up to 237MB after reseting the project.... Is that normal ???
It is 359.01 MB for me, the
)
It is 359.01 MB for me, the second after climateprediction.net with 477.67 MB.
Luckily disk space is cheap.
Tullio
RE: It's now up to 237MB
)
It's perfectly normal, particularly at the current time when the old S5R6 run is being finished up and the new S5GCE run is in full flight. You could quite easily have several times the normal complement of large data files on board and if you wish to contribute to this project, you really need to have these files.
Several weeks ago when you first asked this question, you were advised that resetting the project was a way to remove data files if you were finished with them and all results had been completed and returned. Resetting is NOT a necessary operation and it will NOT give you a permanent reduction in disk usage. Resetting simply throws away all existing programs and data and (if you intend to continue crunching) allows you to download fresh copies of everything. At the end of it, you may very well have a greater disk usage than before, depending on what the scheduler decides to give you in the way of new large data files. If you have finished crunching and wish to withdraw from the project, then you could consider detaching rather than resetting, but if you have any intention to crunch more tasks in the future, you should do neither, as the files you already have on board will continue to be needed.
Bernd also mentioned (because E@H uses locality scheduling) that having 90MB of large data files was not at all unusual for just one task. If you don't know what locality scheduling is, try googling the term. The very top hit will take you to the BOINC documentation about it, which admittedly, is rather technical.
In essence, the E@H project analyses vast amounts of data and a way needed to be found to 'localise' data to groups of tasks. So the scheduler sends you (and a small number of other hosts as well) a particular subset of large data files and then tries very hard to send you all only the tasks that belong to that data. If you have a fast machine dedicated to the project, you can get very large numbers of tasks all belonging to the same data subset. If you have a slow machine or if your machine is shared amongst many other projects, you may find that the other hosts using your subset gobble up all the tasks before you have requested many and you may need to be getting new data subsets in order to get new tasks. It's quite possible to have a few data subsets on the go at any one time and each data subset can be more than 50MB. If you can't afford 250MB of disk space, E@H is really not the project for you.
Your computers are hidden so it's not possible to see what computers you have and how frequently it/they request new work or how many different data subsets might be involved. In any case, with disks being so large and disk space being so cheap these days, do you really need to be concerned about the odd 250MB?
Cheers,
Gary.
RE: Resetting simply throws
)
Actually, and I am not sure when this changed or if it ever changed, resetting will only throw away the executables and data (.dat) files. You'll be surprised how much gunk stays behind in your projects\einstein.phys.uwm.edu\ directory after a controlled No New Tasks followed by a project reset.
It won't throw away all picture files, libraries, or extensionless files. (IOW all the real data files are kept.)
I found that out on a backup of my Data Directory, after I had set all projects to NNT, ran what work I had, uploaded & reported all that and then did a reset on all projects. The total size of the directory was at that time still 270,843,455 bytes (~270MB), 7-zipped it came out at 155,304KB.
Checking the different Projects\ sub-directories, I saw that both Einstein and CPDN (which I hadn't run for over a year) kept all that gunk that I tried to get rid of. Which has resulted in me requesting Trac ticket [trac]#978[/trac]. :-)
(Manually removing everything from every projects\{project}\ directory, I still ended up with a 7-zipped file of 76,940KB (180,146,422 bytes unpacked))