this post was submitted on 19 Jun 2024
713 points (98.8% liked)

Technology

58303 readers
10 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] -2 points 5 months ago (3 children)

Any information on the GPU they are pairing with it?

Does anyone know if it's possible to use a regular AMD or Nvidia GPU with it?

[–] [email protected] 38 points 5 months ago* (last edited 5 months ago) (1 children)

This is not for someone to daily drive. You’ll probably get better performance duct taping and raspberry pi to Bluetooth keyboard and 7 inch pi display.

[–] [email protected] 3 points 5 months ago (1 children)

haha, that doesn't answer the question at all. But I appreciate you.

[–] [email protected] 0 points 5 months ago* (last edited 5 months ago)

It does actually.

Edit: It’s an article about how a company is going to assist in providing RISC 5 dev boards to framework. It’s not about a consumer ready product with a dedicated GPU.

[–] [email protected] 10 points 5 months ago* (last edited 5 months ago)

The GPU inside the processor/soc has the following specifications:

  • Imagination BXE-4-32 GPU with support for OpenCL 1.2, OpenGL ES 3.2, Vulkan 1.2
  • Video Decoder – H.265, H.264 4K @ 60fps or 1080p @ 30fps, MJPEG
  • Video Encoder – H.265/HEVC Encoder, 1080p @ 30fps

I don't think you'll be able to use a separate/external GPU with it. Thunderbolt support is highly unlikely and that processor has only 1 or 2 PCIe lanes (depending how USB is connected), which is likely already used for WiFi.

[–] [email protected] 9 points 5 months ago (1 children)

The processor it's using is linked in the article: https://www.cnx-software.com/2022/08/29/starfive-jh7110-risc-v-processor-specifications/

It's a system-on-chip (SoC) design with an embedded GPU, the Imagination BXE-4-32, which appears to be designed mainly for smart TVs and set-top boxes.

The SoC itself only has two PCIe 2.0 lanes on separate interfaces so you can't use both for the same device, and one is shared with the USB 3.0 interface.

That's not even enough bandwidth to drive an entry-level notebook GPU from over a decade ago. Seriously: the GeForce GT 520M, launched January 2011, wants a full PCIe 2.0 x16 interface. Same with the Raedeon HD 6330M. You could probably get away with just 8 lanes if you had to, but not only one.

The other commenter wasn't kidding by saying you could get more power out of a Raspberry Pi 4. It's even mentioned in the article.

[–] [email protected] 4 points 5 months ago (1 children)

Seriously: the GeForce GT 520M, launched January 2011, wants a full PCIe 2.0 x16 interface. Same with the Raedeon HD 6330M. You could probably get away with just 8 lanes if you had to, but not only one.

Connecting a GPU with just one PCIe lane isn't the biggest problem. You'll just slow down data exchange between the CPU and GPU (mostly loading textures and vertex positions).

If your game mostly relies on shaders and renders lots of rather static stuff, you'll mostly just get longer loading times but FPS shouldn't suffer too much.

[–] [email protected] 3 points 5 months ago (1 children)

Given how much modern games stream data in and out of VRAM, I think it would actually be quite a significant issue. Although, for modern games the 520M would probably be below minimum requirements anyway. It was just to illustrate my point.

[–] [email protected] 2 points 5 months ago

It would be obviously "an issue" and drastically reduce performance in many cases, but compared to the buildin igpu, you'd probably still be able to get a much better performance for lots of applications.