Afaik most LLMs run purely on the GPU, dont they?

So if I have an Nvidia Titan X with 12GB of RAM, could I plug this into my laptop and offload the load?

I am using Fedora, so getting the NVIDIA drivers would be… fun and already probably a dealbreaker (wouldnt want to run proprietary drivers on my daily system).

I know that using ExpressPort adapters people where able to use GPUs externally, and this is possible with thunderbolt too, isnt it?

The question is, how well does this work?

Or would using a small SOC to host a webserver for the interface and do all the computing on the GPU make more sense?

I am curious about the difficulties here, ARM SOC and proprietary drivers? Laptop over USB-c (maybe not thunderbolt?) and a GPU just for the AI tasks…

  • Pantherina@feddit.deOP
    link
    fedilink
    English
    arrow-up
    3
    ·
    8 months ago

    In general I wouldn’t recomend a titan x unless you already have one because it’s been deprecated in cuda, so getting modern libraries to work will be a pain.

    Omg I spent too much on this… thanks for the heads up, that is a major fuckup

    • Possibly linux@lemmy.zip
      link
      fedilink
      English
      arrow-up
      1
      ·
      8 months ago

      AMD cards are slowly starting to get better. Not great yet but if you want a card isn’t Nvidia then it might be for you.