Aussie living in the San Francisco Bay Area.
Coding since 1998.
.NET Foundation member. C# fan
https://d.sb/
Mastodon: @dan@d.sb

  • 0 Posts
  • 188 Comments
Joined 3 years ago
cake
Cake day: June 14th, 2023

help-circle

  • I definitely agree with you!

    I’m using AI a little bit myself, but I’m an experienced developer and fully understand the code it’s writing (and review all of it manually). I use it for tedious things, where I could do it myself but it’d take much longer. I don’t let AI write commit messages or PR descriptions for me.

    At work, I reject AI slop PRs, but it’s becoming harder since AI can submit so much more code than humans can, and there’s people that are less stringent about code quality than I am. A lot of the issues affecting open-source projects are affecting proprietary code too. Amazon recently had to slow down with AI and get senior devs to review AI-written code because it was causing stability issues.












  • QuickSync is more than sufficient for most users. It can handle several concurrent 4K transcode. It’s also not that common to have to transcode, unless you stream your media content when away from home a lot, and have poor upload speed.

    If going Intel, there’s different models of Intel iGPU, so I’d go for the lowest-end GPU that has the higher end iGPU. My home server is a few years old and has an Intel Core i5 13500. The difference between the 13400 and 13500 looks small on paper, but the 13400 only has UHD Graphics 730 while the 13500 had UHD Graphics 770 which can handle double the number of concurrent transcodes.

    Intel iGPUs also support SR-IOV which lets you share one iGPU across multiple VMs. For example, if you have a Plex server on the host Linux system, and Blue Iris in a Windows Server VM, and both need to use hardware transcoding.

    I’ve heard AMD’s onboard graphics are pretty good these days, but I haven’t tried AMD CPUs on a server.