einstein@home: "All-Sky Gravitational Wave search on O3 (O3AS)" extremely inefficient

StefanR5R

Elite Member
Dec 10, 2016
5,591
8,013
136
I haven't been running einstein@home since they stopped "Gamma-ray pulsar binary search #1 on GPUs (FGRPB1G)". So I need to find my way around the other GPU applications now.

At first try, I received O3AS work. It started out using an RTX 4090 by less than 50% SM utilization and less than 150 W board power. I was about to write me an app_config.xml which would launch 2 tasks concurrently per GPU. But the task came to about half of its completion percentage and ceased to use the GPU altogether! It went on to use 100% of one CPU thread instead.

Dammit! How hard are these projects trying to prevent me from keeping my apartment warm?

Trying "Binary Radio Pulsar Search (MeerKAT) (BRP7)" next.
 

StefanR5R

Elite Member
Dec 10, 2016
5,591
8,013
136
BRP7 does use GPU almost all of the time, at least. But its heating characteristics are lacking too.

1 task per GPU: 94 % SM utilization, 215…235 W power draw
2 tasks per GPU: 100 % SM utilization, 210...225 W power draw

…edit, which isn't a lot for an RTX 4090, obviously.
Gonna have to try Petri's BRP7 app.

Edit 2, Petri's app,
1 task per GPU: 100 % SM utilization, 235…255 W power draw
 
Last edited:

mmonnin03

Senior member
Nov 7, 2006
218
221
116
Yes the O3AS app sucks.
BRP7 should only be run with the custom app. 1x is faster than 2x but I haven't seen those CUDA options.
app_config.xml > E@H's site implementation
 

gsrcrxsi

Member
Aug 27, 2022
46
26
51
the O3AS app has gone through a few iterations. I think as of this post, you were already trying the latest one, but for a little background and context i'll try to explain the behavior or best practices.

after the application had been solidified, it was basically running equal time chunks for the GPU and CPU portion. for example, on a single task, a 2080Ti was taking about 8 minutes crunching on the GPU, then another 8 minutes on the CPU sort/search final portion. you could essentially double your throughput by staggering the start of a second task so that the GPU part of the second task started at the CPU portion of the first task. and you might be able to negate the shift over time running more multiples (if you had the VRAM for it, each task took about 4GB)

now, it's a little more complicated. the developers decided to essentially split each task into two halves. instead of running 1x 2Hz searches, they now run 2x 1Hz searches. they did this primarily to lower VRAM use to allow lower end GPUs to be able to contribute, in an effort to increase overall contribution. the tasks now only need about 2GB VRAM. but the behavior is now GPU->CPU->GPU->CPU so the timing of multiples might be more complicated. with something like a 4090 and 24GB VRAM, I'd probably try some high number of multiples like 4-8x with random staggered starts to see if you can combat the low GPU utilization from the CPU-only times, as well as the random drift that will happen leading to undesirable start times and tasks lining up.

but in any case, Petri's BRP7 will probably still give more points, and be easier to manage. 1x without MPS might be fastest (it has been on my cards, but I dont have something as powerful as a 4090). or multiples with MPS. I like 3x with 40% active thread percentage on all my cards.
 

StefanR5R

Elite Member
Dec 10, 2016
5,591
8,013
136
Thanks for the insights! Some time after my post, I too noticed that O3AS isn't just switching from GPU to CPU once, but has got at least (or according to your info, exactly) one more switch back to GPU and another forth to CPU.

My space heater in question consists of two 4090 GPUs which are driven by a 4-core/8-thread Kaby Lake CPU. Throughput gain from HyperThreading isn't very good, but maybe it would be reasonable to run a GPU feeder on one hyperthread and a CPU worker on the paired hyperthread. Given this, the following setup could be investigated:

GPU0, CPU0 ----> task1 using GPU _ CPU _ GPU _ CPU
GPU0, CPU4 ---------------> task2 using GPU _ CPU _ GPU _ CPU
GPU0, CPU1 ----> task3 using GPU _ CPU _ GPU _ CPU
GPU0, CPU5 ---------------> task4 using GPU _ CPU _ GPU _ CPU

GPU1, CPU2 ----> task5 using GPU _ CPU _ GPU _ CPU
GPU1, CPU6 ---------------> task6 using GPU _ CPU _ GPU _ CPU
GPU1, CPU3 ----> task7 using GPU _ CPU _ GPU _ CPU
GPU1, CPU7 ---------------> task8 using GPU _ CPU _ GPU _ CPU

Obviously, timing would be crucial, and depend on a suitable balance of GPU time vs. CPU time within tasks as well as on consistency between different workunits. If necessary, a background script could monitor GPU utilizations and suspend/resume individual tasks.

However, my actual goal on Wednesday was — and still is — to let this particular computer put out a certain heat flux consistently, doing some good science along the way, and accomplish this all with a minimum of hand-holding.

Instead of this dual-GPU space heater, I could have deployed one of my dual-CPU space heaters. However, those were not as easy to move to the other room at this time. I think I should work on making some more of my dual socket computers more mobile than they currently are.
 
Last edited:
Reactions: crashtech

gsrcrxsi

Member
Aug 27, 2022
46
26
51
also the BRP7 app seems to be heavily memory bandwidth limited, which explains why a 4090 isnt massively faster than a 3090/3080ti. it's faster, but not a lot, and not by the amount you'd expect.
 

gsrcrxsi

Member
Aug 27, 2022
46
26
51
However, my actual goal on Wednesday was — and still is — to let this particular computer put out a certain heat flux consistently, doing some good science along the way, and accomplish this all with a minimum of hand-holding.

yeah, I moved one of my systems from the garage to the living space to add heat for the winter and try to reduce NG use.
 

StefanR5R

Elite Member
Dec 10, 2016
5,591
8,013
136
also the BRP7 app seems to be heavily memory bandwidth limited, which explains why a 4090 isnt massively faster than a 3090/3080ti. it's faster, but not a lot, and not by the amount you'd expect.
I never looked into tuning my GPUs beyond reducing their power limits, and in case of air-cooled GPUs, increasing their fan speeds. I wonder if memory clocks on the 4090 could be cranked up from the default, for use cases such as BRP7.
 

gsrcrxsi

Member
Aug 27, 2022
46
26
51
it might help marginally, but watch for memory errors pushing it too far. I pretty much only run the memory clocks at whatever offset gets them to P0 clocks. for Ampere GDDR6X that's +500, for Turing GDDR6 that's +400. not sure how much the 4090 is penalized in P2 vs P0.
 

StefanR5R

Elite Member
Dec 10, 2016
5,591
8,013
136
However, my actual goal on Wednesday was — and still is — to let this particular computer put out a certain heat flux consistently, doing some good science along the way, and accomplish this all with a minimum of hand-holding.
PS, putting out a lot of heat consistently and without any hand-holding at all — using GPUs — would obviously have been accomplished if I had chosen PrimeGrid. In these aspects, number theory handily beats molecular simulations (Folding@Home) or signal processing (Einstein@Home). But (a) I already have quite many PrimeGrid points, (b) none of their conjecture projects use GPUs, (c) their sieving projects have almost accomplished their goals for now (as has the Arithmetic Progressions project), (d) what's left at PrimeGrid for GPUs are the pure Big-Prime-Finding projects which I am not very much in to.
 
sale-70-410-exam    | Exam-200-125-pdf    | we-sale-70-410-exam    | hot-sale-70-410-exam    | Latest-exam-700-603-Dumps    | Dumps-98-363-exams-date    | Certs-200-125-date    | Dumps-300-075-exams-date    | hot-sale-book-C8010-726-book    | Hot-Sale-200-310-Exam    | Exam-Description-200-310-dumps?    | hot-sale-book-200-125-book    | Latest-Updated-300-209-Exam    | Dumps-210-260-exams-date    | Download-200-125-Exam-PDF    | Exam-Description-300-101-dumps    | Certs-300-101-date    | Hot-Sale-300-075-Exam    | Latest-exam-200-125-Dumps    | Exam-Description-200-125-dumps    | Latest-Updated-300-075-Exam    | hot-sale-book-210-260-book    | Dumps-200-901-exams-date    | Certs-200-901-date    | Latest-exam-1Z0-062-Dumps    | Hot-Sale-1Z0-062-Exam    | Certs-CSSLP-date    | 100%-Pass-70-383-Exams    | Latest-JN0-360-real-exam-questions    | 100%-Pass-4A0-100-Real-Exam-Questions    | Dumps-300-135-exams-date    | Passed-200-105-Tech-Exams    | Latest-Updated-200-310-Exam    | Download-300-070-Exam-PDF    | Hot-Sale-JN0-360-Exam    | 100%-Pass-JN0-360-Exams    | 100%-Pass-JN0-360-Real-Exam-Questions    | Dumps-JN0-360-exams-date    | Exam-Description-1Z0-876-dumps    | Latest-exam-1Z0-876-Dumps    | Dumps-HPE0-Y53-exams-date    | 2017-Latest-HPE0-Y53-Exam    | 100%-Pass-HPE0-Y53-Real-Exam-Questions    | Pass-4A0-100-Exam    | Latest-4A0-100-Questions    | Dumps-98-365-exams-date    | 2017-Latest-98-365-Exam    | 100%-Pass-VCS-254-Exams    | 2017-Latest-VCS-273-Exam    | Dumps-200-355-exams-date    | 2017-Latest-300-320-Exam    | Pass-300-101-Exam    | 100%-Pass-300-115-Exams    |
http://www.portvapes.co.uk/    | http://www.portvapes.co.uk/    |