I currently have a Synology 220+ and a couple of VPS’s, and I’m looking to consolidate, while getting out of Synology’s walled garden. I’ve already got a couple of 3.5’s in the Synology, and 4 2.5’s lying around and I’m planning on running a number of docker containers and a couple of vms.

That said, I’ve never built anything before, and basically just went to PCPartPicker, started with the case, and checked 5-stars on each component and went from there. So… how absurd is my build?

PCPartPicker Part List

Type Item Price
CPU AMD Ryzen 5 5600X 3.7 GHz 6-Core Processor $135.00 @ Amazon
CPU Cooler Cooler Master MasterLiquid 360L Core ARGB Liquid CPU Cooler $90.71 @ Amazon
Motherboard MSI MAG B550 TOMAHAWK ATX AM4 Motherboard $165.99 @ B&H
Memory TEAMGROUP T-Force Vulcan Z 16 GB (2 x 8 GB) DDR4-3200 CL16 Memory $26.99 @ Amazon
Storage Seagate IronWolf NAS 8 TB 3.5" 7200 RPM Internal Hard Drive Purchased For $179.00
Storage Seagate IronWolf NAS 8 TB 3.5" 7200 RPM Internal Hard Drive Purchased For $179.00
Storage Seagate IronWolf NAS 8 TB 3.5" 7200 RPM Internal Hard Drive $159.99 @ Adorama
Case Fractal Design Meshify 2 ATX Mid Tower Case $173.89 @ Newegg
Power Supply Corsair RM650 (2023) 650 W 80+ Gold Certified Fully Modular ATX Power Supply $89.99 @ Corsair
Prices include shipping, taxes, rebates, and discounts
Total $1200.56
Generated by PCPartPicker 2025-05-23 19:32 EDT-0400
  • DesolateMood@lemm.ee
    link
    fedilink
    English
    arrow-up
    3
    ·
    21 hours ago

    Dedicated GPUs are obviously going to be more powerful. I’ve never run ai before so maybe someone else can weigh in on the requirements for it, but I can say for sure that an igpu is good enough for jellyfin transcoding. It also depends on your budget, do you want to spend the extra money just for a dedicated GPU?

    If you go igpu route I think that Intel is recommended over AMD, but you should probably do extra research on that before buying

    • yaroto98
      link
      fedilink
      English
      arrow-up
      2
      ·
      19 hours ago

      My amd igpu works just fine for jellyfin. LLMs are a little slow, but that’s to be expected.

      • themadcodger@kbin.earthOP
        link
        fedilink
        arrow-up
        1
        ·
        17 hours ago

        Yeah, I’m not sure if I really want to deal with an llm. It would mostly be for home assistant, so nothing too crazy.

        • yaroto98
          link
          fedilink
          English
          arrow-up
          2
          ·
          8 hours ago

          I have a very similar NAS I built. The Home Assistant usage doesn’t really even move the needle. I’m running around 50 docker containers and chilling at about 10% cpu.

          • themadcodger@kbin.earthOP
            link
            fedilink
            arrow-up
            1
            ·
            5 hours ago

            The LLM for home assistant, or just HA in general doesn’t move the needle? My HA is also pretty low key, but I was considering the idea of running my own small llm to use with HA to get off of OpenAI. My current AI usage is very small, so I wouldn’t need too much on the GPU side I’d imagine, but I don’t know what’s sufficient.

            • yaroto98
              link
              fedilink
              English
              arrow-up
              2
              ·
              3 hours ago

              Just home assistant doesn’t move the needle. The llms hit the igpu hard and my cpu usage spikes to 70-80% when one is thinking.

              But my llms i’m running are ollama and invokeai each with several different models just for fun.