New BOINC project studies machine learning

Message boards : News : New BOINC project studies machine learning
Message board moderation

To post messages, you must log in.

AuthorMessage
Profile David Anderson
Volunteer moderator
Project administrator
Project developer
Avatar

Send message
Joined: 10 Sep 05
Posts: 677
Message 100006 - Posted: 20 Jul 2020, 2:17:57 UTC

Welcome to MLC@Home, a new project dedicated to understanding and explaining machine learning models. It's based at the CORAL lab at the University of Maryland, Baltimore County (UMBC). It supports Linux/amd64 and Windows/x64.
ID: 100006 · Report as offensive
ProDigit

Send message
Joined: 8 Nov 19
Posts: 572
United States
Message 100020 - Posted: 20 Jul 2020, 16:58:46 UTC

System requirements (64 bit CPU only):
Linux
Windows

Future support:
OSX
Linux / arm
Linux / ppc64le

Not supported:
GPU

Too bad, this kind of project could gain so much traction if GPUs were used...
ID: 100020 · Report as offensive
Jim1348

Send message
Joined: 8 Nov 10
Posts: 252
United States
Message 100023 - Posted: 20 Jul 2020, 17:31:00 UTC - in response to Message 100020.  

Too bad, this kind of project could gain so much traction if GPUs were used...

I was wondering about that. You could probably raise that point more effectively than I could on their forum.
Maybe some volunteers could help? It is a nice project.
ID: 100023 · Report as offensive
Profile Dave

Send message
Joined: 28 Jun 10
Posts: 1005
United Kingdom
Message 100025 - Posted: 20 Jul 2020, 17:40:43 UTC - in response to Message 100020.  

System requirements (64 bit CPU only):
Not supported:
GPU

Too bad, this kind of project could gain so much traction if GPUs were used...


If it is similar to CPDN where each calculation depends on the previous one then it is possible GPU won't give much of a boost if any. Of course it may well be able to make use of parallel processing. I am just putting forward a possible reason.
ID: 100025 · Report as offensive
boboviz
Help desk expert

Send message
Joined: 12 Feb 11
Posts: 360
Italy
Message 100032 - Posted: 21 Jul 2020, 7:13:55 UTC - in response to Message 100020.  

Too bad, this kind of project could gain so much traction if GPUs were used...

We are at the beginning of the project...
ID: 100032 · Report as offensive
ProDigit

Send message
Joined: 8 Nov 19
Posts: 572
United States
Message 100086 - Posted: 22 Jul 2020, 18:27:58 UTC - in response to Message 100025.  
Last modified: 22 Jul 2020, 18:33:11 UTC

System requirements (64 bit CPU only):
Not supported:
GPU

Too bad, this kind of project could gain so much traction if GPUs were used...


If it is similar to CPDN where each calculation depends on the previous one then it is possible GPU won't give much of a boost if any. Of course it may well be able to make use of parallel processing. I am just putting forward a possible reason.


Only if the project depends on Double Precision.
Then an RTX2080 @ 300W is about 10x faster than a 15W dual core Celeron CPU at the same frequency.
GPUs can use VRAM to store the result, which they re-read, and re-process through the shaders/cores, without needing to pass through the PCIE bus again.

If a project can run on Single Precision (or 32bit FPP), an RTX 2080 Ti is about 1.000-1.500x faster (in terms of Flops) than a dual core celeron running at the same frequency.

Long WUs may require a lot of VRAM, especially if tailored to the GPU they're feeding.
An RTX 2080 Ti has ~4350 cores and 11GB of VRAM, that means each WU gets less than 2,5MB of VRAM, or there must be some parallel optimizations allowing blocks of cores (shaders) to run simultaneously sharing VRAM memory blocks.

If Half precision (16bit) worked, Nvidia and AMD are working on doubling their half precision process rates, by processing 2 of them in their cores. Rather than running 1 instruction per core, they can fit 2 of them through an algorithm.
Not sure if this process will be implemented in current line of GPUs.
But to give you an idea, current AI runs at Quarter (8 bit) or half(16bit) precision, and uses between 10 to 300 cores.
Can you imagine the processing power if ran through an RTX 2080 Ti (with over 4000 cores)?
Should boost learning algorithms and AI to near real time.
ID: 100086 · Report as offensive
boboviz
Help desk expert

Send message
Joined: 12 Feb 11
Posts: 360
Italy
Message 100164 - Posted: 31 Jul 2020, 7:16:21 UTC - in response to Message 100086.  

The source code of the project is open.
So, if someone wants to help the development....
ID: 100164 · Report as offensive

Message boards : News : New BOINC project studies machine learning

Copyright © 2020 University of California. Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, Version 1.2 or any later version published by the Free Software Foundation.