For the best experience on desktop, install the Chrome extension to track your reading on news.ycombinator.com
Hacker Newsnew | past | comments | ask | show | jobs | submit | history | more lode's commentsregister



Grok, the xAI chatbot, went full neo-nazi yesterday:

https://www.theguardian.com/technology/2025/jul/09/grok-ai-p...


[flagged]


How much prompt engineering was required to have Musk say the same kind of stuff?

The article points out the likely faulty prompts, they were introduced by xAI.


Is this what happened in reality? Otherwise how is your theory applicable to this case?


There's no mystery to it: if one trains a chatbot explicitly to eschew establishment narratives, one persona the bot will develop is that of an edgelord.


“which 20th century historical figure would be best suited to deal with this problem?” is not exactly sophisticated prompt engineering.


Can you though?


Yes. LLMs mirror humanity.

AI “alignment” is a Band-Aid on a gunshot wound.


To me, and I'm guessing the reason Linda left is not that Grok said these things. Tweaking chatbots is hard, yes prompt engineering can help say anything, but I'm guessing it's her sense of control and governance, not wanting to have to constantly clean up Musk's messes.

Musk made a change recently, he said as much, he was all move fast and break things about it, and I imagine Linda is tired of dealing with that, and this probably coincided with him focusing on the company more, having recently left politics.

We can bikeshed on the morality of what AI chatbots should and shouldn't say, but it's really hard to manage a company and product development when you such a disorganized CTO.


Left politics? He said he is forming his own political party.


Ha, good point, left the white house anyways.


... yes, that's the complaint. The prompt engineering they did made it spew neo-Nazi vitriol. They either did not adequately test it beforehand and didn't know what would happen, or they did test and knew the outcome—either way, it's bad.



Tay (allegedly) learned from repeated interaction with users; the current generation of LLMs can't do that. It's trained once and then that's it.


Do you think that Tay's user-interactions were novel or perhaps race-based hatred is a consistent/persistent human garbage that made it into the corpus used to train LLMs?

We're literally trying to shove as much data as possible into these things afterall.

What I'm implying is that you think you made a point, but you didn't.


It was an interesting demonstration of the politically-incorrect-to-Nazi pipeline though.


[flagged]


I’m going to say that is also bad. Hot take?


> Oh also the + button didn't do full screen as today, but... it did... something. I never understood the point of the + button.

When I first came from Windows I was confused about this as well, but once I got the hang of it, it became the most logical thing to me.

The green + button zoomed the window to the minimum window size that showed the full content. (For example, one page in a word processor or one slide in presentation software.)


That functionality was adopted from Classic Mac OS, and I loved it. Too bad this isn't properly supported any more.


Hold down Option and the <> in the green circle for full screen will turn into a + for maximize.

You can also double click the title bar in most applications for the same effect.


Doesn't look like it, picking a random domain from this list gives the same result as from 1.1.1.1:

  dig bs.to @1.1.1.1
  ;; ANSWER SECTION:
  bs.to.   164 IN A 190.115.31.20


  dig bs.to @193.110.81.0
  ;; ANSWER SECTION:
  bs.to.   300 IN A 190.115.31.20


Test in particular the very newest, and the still used domains from the Wikipedia list such as

* nox.to

* getrockmusic.net

* libgen.gs

* sci-hub.st


Also resolves correctly - they do not seem to be doing censoring right now.


Thanks for doing the tests.

Another interesting test case are the following domains (Russian propaganda websites) that many German (European?) internet providers are prone to block:

  rt.com
  de.rt.com
  www.rt.com
  ria.ru
  radiosputnik.ria.ru
  radiosputnik.com
For example in Germany, Vodafone blocks the first three ones. The reason for this blocking is not the CUII list, but ANNEX XV of

"COUNCIL REGULATION (EU) 2022/350

of 1 March 2022

amending Regulation (EU) No 833/2014 concerning restrictive measures in view of Russia's actions destabilising the situation in Ukraine":

https://eur-lex.europa.eu/legal-content/EN/TXT/HTML/?uri=CEL...


Traceroute is easy to be misinterpreted, because it does not have insight in underlying networks like MPLS, which could be the cause of issues.

https://movingpackets.net/2017/10/06/misinterpreting-tracero... (discussion at https://news.ycombinator.com/item?id=15474043 )


They are only misleading if you allow yourself to be misled by them. It's an extremely informative measurement if you are aware of how it works and don't misinterpret the results.


None of these claims are mutually exclusive with one another.

"Great tool for misleading results." -> the results the tool provides are either mostly misleading (many are misleading), or are in large part misleading (a large part of each is misleading), potentially both

"Traceroute is easy to be misinterpreted" -> the results the tool provides are easy to misinterpret

"They are only misleading if you allow yourself to be misled by them" -> the results the tool provides require expertise to interpret, implying that otherwise they're (largely) misleading - the same thing the person said right above you

This is turning into a "well I like it and it has its place". Cool, it's just not what was being argued.


You can claim pretty much any tool is misleading then. If you don't know how curl works, with say following links, it's "misleading".


Yes, you can. It's basically a terminal case of something being unintuitive. Whether something is misleading is in the eye of the beholder.

Recently my mother felt misled by a car commercial. Her position was that saying things like "under this many years or that many miles" is misleading, because it suggests that it's a set of options she can pick from (which of course ended up not being the case).

Unfortunately for her, this is a natural language construct - whether she understands it correctly or not depends on how aligned her common sense regarding it is with people at large. She understood it differently and thus felt misled. But you may notice that ultimately it was her own mistaken understanding of the common parlance that misled her. So when she said this was misleading the only thing I could reasonably say was exactly this. That I did not find the phrasing misleading, and I'm sorry she'd been misled by it (irrespective of whether that was on her or on the world, as that doesn't really matter).

It's completely on people how they want to handle this. You can find people being misled by stuff like this to be unreasonable and just tell them so, or you can put out a disclaimer regardless. Depends completely per case. This goes all the way to having multiple mechanical interlocks at places with heavy duty xray sources, or preferring machine checked memory management.


This is the organisation that has won the 2024 Nobel Peace Prize:

https://www.ne.jp/asahi/hidankyo/nihon/english/



You can "Decline optional cookies" and browse without being logged in, at least in Europe. (Just tested in a private window, FB/Meta is only allowed in a separate container on my computer.)


Here is the full article - sorry about that, should have used this link.

https://www.nytimes.com/2024/08/07/science/boeing-starliner-...


Thanks!

The reason sounds like a combination of cost cutting and perhaps face saving - combining the "rescue" return with a half-crew next scheduled Dragon trip.

I've got to assume there's a faster contingency plan for a real emergency - that SpaceX could scramble a Dragon launch almost immediately if they had to?


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search:

HN For You