this post was submitted on 30 Dec 2023
25 points (93.1% liked)
homelab
6589 readers
7 users here now
founded 4 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Not possible AFAIK, plus it will degrade the performance due to the latency etc, IMO it’s not feasible and not the best way if you want to leverage your GPU’s horsepower.
You will need to keep the transcoding in the storage server, maybe the rest (a viewer, manager etc) you can move to the Turing Pi 2.
But then, if it’s for a real time decoding, it’s not possible. Rather than getting an SBC like the Pi styled computers, consider getting something like a motherboard that has built in J4125 from Biostar which has a PCI-E slot and move your GPU to that Biostar mobo to handle all your media needs. And keep the storage server GPU-less.
Transcoding on download seems like the easiest use case, I could use Tdarr and have one node with the GPU. But for apps like Immich that use the GPU for both transcoding (raw to jpg?) and for ML purposes (facial recognition) I'm guessing the container will have to run on the hardware where the GPU is, which means Plex and Jellyfin will also have to follow the gpu.
I've definitely thought about moving the GPU to a dedicated mini itx box. Wonder if I could find something rack friendly..