

Because most people do not understand what this technology is, and attribute far too much control over the generated text to the creators. If Copilot generates the text “Trans people don’t exist”, and Microsoft doesn’t immediately address it, a huge portion of people will understand that to mean “Microsoft doesn’t think trans people exist”.
Insert whatever other politically incorrect or harmful statement you prefer.
Those sorts of problems aren’t easily fixable without manual blocks. You can train the models with a “value” system where they censor themselves but that still will be imperfect and they can still generate politically incorrect text.
IIRC some providers support 2 separate endpoints where one is raw access to the model without filtering and one is with filtering and censoring. Copilot, as a heavily branded end user product, obviously needs to be filtered.
I can’t help, just chiming in to say that I’ve also had that experience with Immich. It’s the one service I’ve used that has somehow managed to break itself multiple times like this.
No idea how it happens, I don’t do anything weird with the setup and it just breaks. I’d heard that feedback from other people too but didn’t believe it until it happened to me. It’s been a few months so maybe I’ll try again, I’m just not too happy importing hundreds of gigs of photos multiple times.
So yea just… you’re not alone, good luck.