this post was submitted on 16 Jan 2024
14 points (88.9% liked)

Technology

35125 readers
76 users here now

This is the official technology community of Lemmy.ml for all news related to creation and use of technology, and to facilitate civil, meaningful discussion around it.


Ask in DM before posting product reviews or ads. All such posts otherwise are subject to removal.


Rules:

1: All Lemmy rules apply

2: Do not post low effort posts

3: NEVER post naziped*gore stuff

4: Always post article URLs or their archived version URLs as sources, NOT screenshots. Help the blind users.

5: personal rants of Big Tech CEOs like Elon Musk are unwelcome (does not include posts about their companies affecting wide range of people)

6: no advertisement posts unless verified as legitimate and non-exploitative/non-consumerist

7: crypto related posts, unless essential, are disallowed

founded 5 years ago
MODERATORS
top 2 comments
sorted by: hot top controversial new old
[–] techwizrd 6 points 11 months ago

As a federated learning researcher, I love to see articles introducing the public to the idea. But this article is really drawing a comparison between the fediverse and federated learning that doesn't make sense (to me).

Beyond the fact that data and compute are stored on separate servers, they really aren't similar. Federated learning avoids sharing data by sharing gradients or model updates with a central aggregator; raw data does not leave your device. The fediverse enables easy sharing of data between servers and avoids a central server.

Additionally, this article makes it seem like medical researchers were inspired by the fediverse, but the FedAvg paper was released in 2016—two years before ActivityPub was introduced in 2018.

[–] [email protected] 4 points 11 months ago

This is the best summary I could come up with:


Putting their AI algorithms on special servers within hospitals for "federated learning" could keep privacy standards high while letting researchers unravel new ways to detect and treat diseases.

“The use of AI is just exploding in all facets of life,” said Ronald M. Summers of the National Institutes of Health Clinical Center in Maryland, who uses the method in his radiology research.

Sending patient information outside of a hospital’s firewall is a big risk, so getting permission can be a long and legally complicated process.

Typically, doctors first identify eligible patients for a study, select any clinical data they need for training, confirm its accuracy, and then organize it on a local database.

Once the software receives instructions from the researchers, it can work its AI magic, training itself with the hospital’s local data to find specific disease trends.

It also means machine/AI learning can reach its full potential by training on real-world data so researchers get less biased results that are more likely to be sensitive to niche diseases.


The original article contains 667 words, the summary contains 170 words. Saved 75%. I'm a bot and I'm open source!