There's no mystery to it: if one trains a chatbot explicitly to eschew establishment narratives, one persona the bot will develop is that of an edgelord.
To me, and I'm guessing the reason Linda left is not that Grok said these things. Tweaking chatbots is hard, yes prompt engineering can help say anything, but I'm guessing it's her sense of control and governance, not wanting to have to constantly clean up Musk's messes.
Musk made a change recently, he said as much, he was all move fast and break things about it, and I imagine Linda is tired of dealing with that, and this probably coincided with him focusing on the company more, having recently left politics.
We can bikeshed on the morality of what AI chatbots should and shouldn't say, but it's really hard to manage a company and product development when you such a disorganized CTO.
... yes, that's the complaint. The prompt engineering they did made it spew neo-Nazi vitriol. They either did not adequately test it beforehand and didn't know what would happen, or they did test and knew the outcome—either way, it's bad.
Do you think that Tay's user-interactions were novel or perhaps race-based hatred is a consistent/persistent human garbage that made it into the corpus used to train LLMs?
We're literally trying to shove as much data as possible into these things afterall.
What I'm implying is that you think you made a point, but you didn't.
> Oh also the + button didn't do full screen as today, but... it did... something. I never understood the point of the + button.
When I first came from Windows I was confused about this as well, but once I got the hang of it, it became the most logical thing to me.
The green + button zoomed the window to the minimum window size that showed the full content. (For example, one page in a word processor or one slide in presentation software.)
Another interesting test case are the following domains (Russian propaganda websites) that many German (European?) internet providers are prone to block:
They are only misleading if you allow yourself to be misled by them. It's an extremely informative measurement if you are aware of how it works and don't misinterpret the results.
None of these claims are mutually exclusive with one another.
"Great tool for misleading results." -> the results the tool provides are either mostly misleading (many are misleading), or are in large part misleading (a large part of each is misleading), potentially both
"Traceroute is easy to be misinterpreted" -> the results the tool provides are easy to misinterpret
"They are only misleading if you allow yourself to be misled by them" -> the results the tool provides require expertise to interpret, implying that otherwise they're (largely) misleading - the same thing the person said right above you
This is turning into a "well I like it and it has its place". Cool, it's just not what was being argued.
Yes, you can. It's basically a terminal case of something being unintuitive. Whether something is misleading is in the eye of the beholder.
Recently my mother felt misled by a car commercial. Her position was that saying things like "under this many years or that many miles" is misleading, because it suggests that it's a set of options she can pick from (which of course ended up not being the case).
Unfortunately for her, this is a natural language construct - whether she understands it correctly or not depends on how aligned her common sense regarding it is with people at large. She understood it differently and thus felt misled. But you may notice that ultimately it was her own mistaken understanding of the common parlance that misled her. So when she said this was misleading the only thing I could reasonably say was exactly this. That I did not find the phrasing misleading, and I'm sorry she'd been misled by it (irrespective of whether that was on her or on the world, as that doesn't really matter).
It's completely on people how they want to handle this. You can find people being misled by stuff like this to be unreasonable and just tell them so, or you can put out a disclaimer regardless. Depends completely per case. This goes all the way to having multiple mechanical interlocks at places with heavy duty xray sources, or preferring machine checked memory management.
You can "Decline optional cookies" and browse without being logged in, at least in Europe. (Just tested in a private window, FB/Meta is only allowed in a separate container on my computer.)
The reason sounds like a combination of cost cutting and perhaps face saving - combining the "rescue" return with a half-crew next scheduled Dragon trip.
I've got to assume there's a faster contingency plan for a real emergency - that SpaceX could scramble a Dragon launch almost immediately if they had to?