This may not be factually wrong but it’s not well written, and probably not written by a person with a good understanding of how Gen AI LLM’S actually work. This is an algorithm that generates the next most likely word or words based on its training data set using math. It doesn’t think. It doesn’t understand. It doesn’t have dopamine receptors in order to “feel”. It can’t view “feedback” in a positive or negative way.
Now that I’ve gotten that out of the way, it is possible that what is happening here is that they trained the LLM on a data set that has a less than center bias. If it responds to a query with something generated statistically from that data set, and the people who own the LLM don’t want it to respond with that particular response they will add a guardrail to prevent it from using that response again. But if they don’t remove that information from the data set and retrain the model, then that bias may still show up in responses in other ways. And I think that’s what we’re seeing here.
You can’t train a Harry Potter LLM on both the Harry Potter Books and Movies and the Harry Potter online fanfiction available and then tell it not to respond to questions about canon with fanfiction info if you don’t either separate and quarantine that fanfiction info, or remove it and retrain the LLM on a more curated data set.
I will get stuff because that’s what my parents own. They don’t have large amounts of liquid cash but my dad owns his house and my mom owns lots of antique furniture (passed down from her family) and jewelry (she has a problem with buying shiny gold and silver pieces). But there’s also 8 of us kids so the likelihood is that we each won’t get much in the way of any real inheritance even from what they do have.
It’s easier for most everyone involved to just let them live out their lives using what they have earned along the way. So I told my parents pretty much the same thing. Take care of yourselves. We’ll be alright.