I currently have a Synology 220+ and a couple of VPS’s, and I’m looking to consolidate, while getting out of Synology’s walled garden. I’ve already got a couple of 3.5’s in the Synology, and 4 2.5’s lying around and I’m planning on running a number of docker containers and a couple of vms.

That said, I’ve never built anything before, and basically just went to PCPartPicker, started with the case, and checked 5-stars on each component and went from there. So… how absurd is my build?

PCPartPicker Part List

Type Item Price
CPU AMD Ryzen 5 5600X 3.7 GHz 6-Core Processor $135.00 @ Amazon
CPU Cooler Cooler Master MasterLiquid 360L Core ARGB Liquid CPU Cooler $90.71 @ Amazon
Motherboard MSI MAG B550 TOMAHAWK ATX AM4 Motherboard $165.99 @ B&H
Memory TEAMGROUP T-Force Vulcan Z 16 GB (2 x 8 GB) DDR4-3200 CL16 Memory $26.99 @ Amazon
Storage Seagate IronWolf NAS 8 TB 3.5" 7200 RPM Internal Hard Drive Purchased For $179.00
Storage Seagate IronWolf NAS 8 TB 3.5" 7200 RPM Internal Hard Drive Purchased For $179.00
Storage Seagate IronWolf NAS 8 TB 3.5" 7200 RPM Internal Hard Drive $159.99 @ Adorama
Case Fractal Design Meshify 2 ATX Mid Tower Case $173.89 @ Newegg
Power Supply Corsair RM650 (2023) 650 W 80+ Gold Certified Fully Modular ATX Power Supply $89.99 @ Corsair
Prices include shipping, taxes, rebates, and discounts
Total $1200.56
Generated by PCPartPicker 2025-05-23 19:32 EDT-0400
  • @DesolateMood@lemm.ee
    link
    fedilink
    English
    39 days ago

    Dedicated GPUs are obviously going to be more powerful. I’ve never run ai before so maybe someone else can weigh in on the requirements for it, but I can say for sure that an igpu is good enough for jellyfin transcoding. It also depends on your budget, do you want to spend the extra money just for a dedicated GPU?

    If you go igpu route I think that Intel is recommended over AMD, but you should probably do extra research on that before buying

    • @yaroto98@lemmy.org
      link
      fedilink
      English
      28 days ago

      My amd igpu works just fine for jellyfin. LLMs are a little slow, but that’s to be expected.

      • themadcodgerOP
        link
        fedilink
        18 days ago

        Yeah, I’m not sure if I really want to deal with an llm. It would mostly be for home assistant, so nothing too crazy.

        • @yaroto98@lemmy.org
          link
          fedilink
          English
          28 days ago

          I have a very similar NAS I built. The Home Assistant usage doesn’t really even move the needle. I’m running around 50 docker containers and chilling at about 10% cpu.

          • themadcodgerOP
            link
            fedilink
            18 days ago

            The LLM for home assistant, or just HA in general doesn’t move the needle? My HA is also pretty low key, but I was considering the idea of running my own small llm to use with HA to get off of OpenAI. My current AI usage is very small, so I wouldn’t need too much on the GPU side I’d imagine, but I don’t know what’s sufficient.

            • @yaroto98@lemmy.org
              link
              fedilink
              English
              28 days ago

              Just home assistant doesn’t move the needle. The llms hit the igpu hard and my cpu usage spikes to 70-80% when one is thinking.

              But my llms i’m running are ollama and invokeai each with several different models just for fun.