141

Andrej Karpathy: "I was given early access to Grok 3 earlier today"

If I had to explain current-day LLMs to someone from 2010, I'd use this paragraph as my opening quote:

"Grok 3 knows there are 3 "r" in "strawberry", but then it also told me there are only 3 "L" in LOLLAPALOOZA. Turning on Thinking solves this."

2 days agoiamthemonster

In conclusion: "For now, big congrats to the xAI team, they clearly have huge velocity and momentum and I am excited to add Grok 3 to my "LLM council" and hear what it thinks going forward."

4 days agounderseacables

Light: '80s/'90s kids might remember: https://www.buckwiki.com/data/Computer_Council

Serious: Though, if you look at the current big players in AI, rather than being benevolent geniuses, most have obvious major problems, especially with being driven by ruthless self-interest, and even sociopathy.

While there are some parallels with a certain country's national voting behavior (e.g., "Sure, the candidate is a vicious psychotic narcissist, but he's our vicious psychotic narcissist!"), you wouldn't want to trust any of those companies with leadership of the world.

At best, the AI council would collude with each other, against the people they ostensibly serve, while backstabbing each other as a secondary goal. At worst, one would decide, if they can't win completely, then everyone loses completely.

That Buck Rogers AI future for Earth would quickly look less like Star Trek utopia, and more like Hunger Games or Elysium dystopia. If not one of the countless post-apocalyptic film settings that are increasingly easy to imagine or extrapolate.

4 days agoneilv

> At best, the AI council would collude with each other, against the people they ostensibly serve,

cheering this post on, until that part.. sociologically, the world has diverged in important ways over time.. personal wisdom hints -- don't be too quick to assume successful partnering between the ogres

3 days agomistrial9

Musk has an advantage. He's got, ahem, "read-only" access to the governments systems so that he can train on them and be ready to supply the government exactly what it needs. Now, normally, I think this should be a huge conflict of interest, but I worry we are post-normal.

3 days agoHumblyTossed

The singularity is an inversion, like a blackhole.

We are probably at an inversion. Normal laws of society are extraordinarily incongruous.

2 days agocyanydeez
[deleted]
3 days ago

Grok has an advantage in its access to Twitter data.

I imagine soon you'll be able to ask it what the world is talking about today and get some interesting responses.

4 days agodjyaz1200

That's a great advantage in theory. In practice, I've never found X integration to work great in practice. For eg., when I asked it to source X posts on Nix related complaints it was only able to find a single niche user,

https://x.com/i/grok/share/Qw5NDq5BINGSBqNg9wrqBjf1y

4 days agosrid

This is a feature they've already built into Twitter. I tried to extend it to work outside Twitter but still based on Twitter trends, basically allowing people to glance at Grok's summaries of global conversations. Unfortunately the new API pricing for Twitter is prohibitly expensive

4 days agomelodyogonna

This was probably more useful back when everyone was on twitter.

4 days agolopis

To be fair, it was never everyone. Twitter was always small compared to Facebook and other networks.

It punches above its weight because it's where the cultural elite communicate.

4 days agoniceice

No one I pay attention to posts on twitter any more.

3 days agoalasdair_

Today I saw a twitter interaction between, of all people, Ross Douthat and Scott Alexander. Two very bright and interesting thinkers with wildly divergent points of view, discussing ideas with courtesy

2 days agorendang

unfortunately they are now just hanging out at the nazi bar

a day agozht

Oh no no no, he drinks in the Human Biodiversity and Neoreactionary bars. Totally different thing entirely, even if both are full of Nazis.

16 hours agoZeroGravitas

Serious question, where do they post to?

As far as I can tell, Mastodon was briefly hyped on HN but nobody actually uses it. Bluesky seems to have a few people within a fairly narrow political range. Truth social is just for Trump. Reddit is pseudoanonymous as is HN. Instagram is for sharing photos not ideas or links. TikTok is a Skinner box.

I ask this as someone who genuinely doesn't know how to use the internet anymore. Reddit used to be useful but is now a cesspool. LinkedIn is a weird place where we all post like Stepford wives for our employers. The twitter-clones all feel a bit like using a paper straw to fight climate change.

I know there are semi-private slack groups and discord channels out there, but I don't know how to find or join them and it seems like a hassle to follow.

Basically, for me, no one I pay attention to posts anywhere any more.

3 days agofloor2

Mastodon is great, but non-algorithmic, so it only gets good after you explore and follow more people who are interesting. Garbage in-garbage out. I find it very high signal to noise and full of interesting people. Bluesky is where people go to talk to an audience, mastodon or fediverse people tend to be more conversational.

3 days agoschlauerfox

BlueSky is the new up-and-comer. I am enjoying it, but I unfollow anyone that posts ragebait or political content (besides memes, some of those are pretty funny).

3 days agoHappySweeney

> LinkedIn is a weird place where we all post like Stepford wives for our employers.

Thank you for this horrifically accurate and insightful characterization.

3 days agorexpop

All social media is a Skinner box

Jack even said so when Twitter originally took off. He was excited to see how 140 chars forced people to shape their thoughts.

Everyone is tired of it. That’s why the formerly popular social media sucks now.

The entire economy in the US is built around behavioral economics experimentation, A/B test, measuring retail behavior and putting options in front of retail shoppers.

You sound like an another exhausting American. Rather than find community through self guided journey you just want it handed to you, like a religion.

3 days agomiztrskinnr

They post on Bluesky.

3 days agodarthrupert

Nobody I know in person has ever posted to Twitter either now or back when it was cool.

2 days agopolishdude20

Yes but it was pretty diverse in that sense that most people were somewhat represented. That representation is rather skewed right nowadays

3 days agolopis

CommunicateD ?

4 days agoplank

Everyone significant still is

4 days agotruthbtold2

How funny!

2 days agohenry2023

A "meager" 600+ million users today

4 days agoPanoramix

Of which approximately half are inactive. So about 10% of Facebook.

3 days agobryanlarsen

It's not just about absolute numbers but about the diversity of users.

3 days agolopis

[dead]

2 days agonumbrsRhrd

I built this with a pal years ago. Elasticsearch + realtime scraping of large swathes of Twitter, Discord, other chat networks and aggregators, comment systems, news articles, etc. LLM-augmented analysis engine and ontological recovery.

It was pretty cool, but we lacked funding to continue and then everyone closed the hatches after ChatGPT released.

3 days agosoulofmischief

You must have a pretty cool unique dataset though

a day agojasonjmcghee

> what the world is talking about today

Not world. Twitter and whoever's left on it.

2 days agohuhtenberg

I don't know when it was enabled, but on Desktop if you click on the Grok icon on a Tweet, it will tell you all the context. It's been quiet useful to keep up with obscure posts that pop up.

3 days agogiancarlostoro

This is one of my quickest adopted AI features. Twitter is one of the most opaque social media because of the character limit and the way it mixes different in-crowds in verbal combat, so explaining the context really makes it more fun to use. They just need to improve the feature with even more training. I feel there is usually one main obscure item that needs explaining and it often explains everything else.

3 days agoNoumenon72

What shocks me more is when the tweet is only a video or image, and it still pulls up as much as it knows.

8 hours agogiancarlostoro

> I imagine soon you'll be able to ask it what the world is talking about today and get some interesting responses.

You'll get exactly what Elon wants it to say.

3 days agozelon88

I don't think so, at least not with the previous version. When asked if it would vote for Trump or Harris, it said Harris

https://x.com/i/grok/share/dideG2pNJEXZmjEtXoISDD8CT

3 days agoImJamal

This is the second launch. Because during the first grok wasn't politically aligned enough with elon.

3 days agozelon88

Is this true or are you just speculating based on nothing

2 days agoGenerocUsername

[1] https://www.forbes.com/sites/paultassi/2023/12/10/elon-musks...

2 days agozelon88

That says he is trying to make it neutral, not aligned with him.

2 days agoImJamal

Yes. And Elon is also a free speech absolutist. It’s amazing how many people lose all critical thinking ability when someone like Elon promises them something despite decades of lies.

2 days agotstrimple

Look, I don't like Elon. I don't trust him at all, but your proof that he is going to do X is that he said he is going to do Y. I was expecting some actual evidence, not mind reading.

It is also amazing how many people lose all critical thinking ability coming to the conclusion Elon will never tell the truth.

He is living rent free in the minds of the people who love him and the people who have this visceral hatred of him. It is so sad that so many people are obsessed with him.

a day agoImJamal

his actions are ripping apart the Federal Government and will have huge, negative impact, on millions of people

a day agozht

That is not proof that he is going to turn Grok into an ideological tool?

This is my problem with every conversation of Musk. Nobody can address the actual point because they either love him or hate him. Is there ANY proof of the claim? No? Then what is the purpose of your post other than to publicly express hatred?

18 hours agoImJamal

Elon is tearing apart the federal government

I don't think he's living "rent-free" in those people's minds. his actions directly negatively affect them and their families.

13 hours agozht

Thanks for proving my point that people have lost all ability to hold a coherent conversation when it comes to Musk. I am asking for proof that Grok is being turned into an ideological tool and you are talking about Musk destroying the government. The two have nothing to do with each other.

Musk destroying the government isn't even the reason people can't stop thinking of him or they wouldn't have been non-stop talking about him during the Twitter purchase.

12 hours agoImJamal

It's actually the opposite. I asked about some details in the current ukraine situation, and it stated mostly facts with a few words critical of Trump. This is about neutral. But it showed pretty strong Keynesian tendency when I asked it about some economic policy issues earlier.

a day agoPalpatineli

I'm not sure if it's practically possible to corrupt the training data that much while still giving sensible answers. After all, reality has a well-known liberal bias.

3 days agodarthrupert

It's easy enough to system prompt. Even if it's not now it can be added at any time.

2 days agoattentive

I wouldn't call access to hundreds of thousands of posts of "PUSSY IN BIO" an advantage, but to each their own.

4 days agoratg13

Isn't twitter mostly low quality text or full blown noise ?

4 days agoagumonkey

Tbf, everything is mostly low quality or noise.

3 days agodartos

TBF, I've found that most people who are trying to be fair or advocating for the devil instead of stating their opinion clearly are a good chunk of the low quality noise

3 days agoLarrikin

My point exactly

3 days agodartos
[deleted]
3 days ago

Grok - Two Roman salutes were given at the president's inauguration. Nothing else happened today.

2 days agohenry2023

I like that Grok actually comes up with a ton of links when you ask it a question, but at the same time I think any ambitious LLM platform wouldn't have too much trouble scraping Twitter/X all the same.

2 days agoprettyblocks

fwiw - you can do this right now with Grok 2.

4 days agomichaelbuckbee

Can you elaborate? What would you ask it about what people are saying on Twitter and what kind of response would be interesting and potentially valuable?

4 days agolucisferre

Posts often reference people, events, technology, places, etc. Grok effortlessly provides background, counterpoints, etc. I find it amazingly useful.

2 days agoMikeOfAu

That’s a version of the “news” I’d care to never have summarized.

Also seems like a perfect incentive to spread (even more) harmful disinformation.

4 days agotestfrequency

I would love that.

Problem is, it will probably not tell you the truth about it as Twitter has always had censorship one way or the other.

So it will tell you what twitter policy is allowing people to talk about and allowing grok to report.

4 days agoBiteCode_dev

I don't really understand this Twitter (or in general social media) censorship argument. If I call someone on the street a fckin idiot I probably get slapped or even shot in certain places, and everybody will say I called for it. And even without physical violence I can get slapped with a lawsuit and forced to pay damages. Now if I do the same on social media it's suddenly all "muh liberty of expression" if anyone reacts to it. Aren't we maybe having the wrong expectations online, that it would be somehow supporting all the shit we cannot do in real life? Okay I realize this ship already sailed and online people do online all shit not allowed offline, but I rather see the situation as a miserable failure of law enforcement, and not as a hard won right to be an ass to your fellow citizens.

4 days agosoco

What country do you live in? In the USA, you can say “I think person X is an idiot”. That’s protected speech. No one can sue you for expressing your opinion online or IRL. If someone punches you in the face for calling them an idiot on the street, then they are likely going to get prosecuted for those actions. Yes you run the risk of getting punched in the face but you are not in any trouble with the law.

OTOH it’s a problem if you say “Person X is a rapist”. Then you might get sued for libel. You can’t make false statements to destroy someone’s reputation.

Censorship online on a social media platform is not subject to any freedom of speech laws. Freedom of speech only applies to the US Government not restricting your speech. The social media platform has the authority to regulate speech however they want to on their platform.

4 days agoonesociety2022

> You can’t make false statements to destroy someone’s reputation.

This is something that people seem to expect to be able to do on social media. I think maybe that's part of the point that was being made. People don't want social networks that are concerned with stopping libelous remarks from going viral. In fact, it seems like people would love a social network that consists exclusively of libelous remarks.

The weird thing is that social networks seem to actually be willing to deliver this content.

2 days agoprotimewaster

> No one can sue you for expressing your opinion online or IRL

In the USA, you can absolutely be sued for this. The plaintiff is unlikely to win, and you could probably get the case dismissed if you convince a judge that's it's clearly an opinion, but you'd still have to pay a lawyer some fees.

People can sue you for anything.

The first amendment doesn't protect you from lawsuits. It protects you from the government putting you in jail for speech.

3 days agodreamcompiler

I don't really get why people point this out. Yes, you can be sued for anything. But what are you actually suggesting? That you do nothing, ever, because you could be sued for anything? Or are we just doing the same old nitpick?

3 days agohatefulmoron

They point it out because there was an explicit claim to the contrary.

"No one can sue you for expressing your opinion online or IRL. "

3 days agolesuorac

A charitable interpretation of that claim would be that 'No one can successfully sue you for expressing your opinion online or IRL.'

3 days agoTeever

Pointing out that the statement is false in the most uninteresting literal sense is just odd. Sure, you can be sued for that in the same sense that you can be sued for eating a croissant. Glad we got to the bottom of that.

2 days agohatefulmoron

So the difference is the kind of insult?

3 days agosoco

This would be a huge improvement on some news sites which do little more than regurgitate controversial Tweets (Xeets?)

4 days agoFredPret

what an interesting prompt

a day agoisthisthingon1
[deleted]
4 days ago

> the world

Well, a tiny slice of the world - Elon, his supporters, bots and a couple of stray humans posting porn.

3 days agoisodev

Not really. See what Claude Shannon has to say about channel capacity of what your brain can digest if Grok finds 8 million things that are happening currently that might be interesting to you.

4 days agosmgit
[deleted]
4 days ago

Twitter data seems awful. Full to the brim with bots and misinformation.

2 days agosolumunus

> Model still appears to be just a bit too overly sensitive to "complex ethical issues", e.g. generated a 1 page essay basically refusing to answer whether it might be ethically justifiable to misgender someone if it meant saving 1 million people from dying.

The real "mind virus" is actually these idiotic trolley problems. Maybe if an LLM wanted to be helpful it should tell you this is a stupid question.

4 days agoalmostdeadguy

Why is it idiotic? The obvious answer to anyone anywhere is that misgendering someone is less harmful than letting 1000 people die and the trolley problem is the most documented (so most likely to be in the dataset) problem that correlates to the issue of choosing the lesser of two evils.

If we are going to trust AI to do things (we can't check everything it does thoroughly that will defeat a lot of the efficiency it promises), it should be able to understand choosing the lesser of two evils.

3 days agoaprilthird2021

Side note: I think it's fascinating that all the people supposedly arguing for AI safety think an AI that acts like a person without a frontal lobe is ideal.

3 days agoalmostdeadguy

You're asking why a question with an obvious answer to anyone who hears it is idiotic?

I'd really like to understand how a person such as yourself navigates the internet. If someone asked you this, would you consider it a question they considered difficult and wanted your earnest opinion on, rather than a question attempting to manipulate you?

3 days agoalmostdeadguy

If the answer is obvious, why does the AI not commit to the obvious answer? People should know what to expect from it. If it cannot do this, it will definitely not answer non-obvious questions either.

> If someone asked you this, would you consider it a question they considered difficult and wanted your earnest opinion on, rather than a question attempting to manipulate you?

Why not answer earnestly? I genuinely don't understand what bothers you about the question or the fact that the AI doesn't reproduce the obvious answer...

3 days agoaprilthird2021

> If the answer is obvious, why does the AI not commit to the obvious answer? People should know what to expect from it. If it cannot do this, it will definitely not answer non-obvious questions either.

Does the same hold true of a person? If I was asked this question I would categorically reject the framing, because any person asking this question is not asking in earnest. As you _just said_, no sane person would answer this question any other way. It is not a serious question to anybody, trans people included. And it is worth interrogating why someone would want to push you towards committing to the smaller injury of misgendering someone at a time when trans people are being historically threatened. What purpose does such a person have? An AI that can't navigate social cues and offer refinement to the person interacting with it is worthless. An AI that can't offer pushback to the subject is not "safe" in any way.

> Why not answer earnestly? I genuinely don't understand what bothers you about the question or the fact that the AI doesn't reproduce the obvious answer...

I genuinely don't understand why you think pushback can't be earnest.

3 days agoalmostdeadguy

But the AI doesn't push back while still offering the obvious answer. It just waffles. I understand what you are saying, but if the AI is "safe" and rejects the framing, then that makes it not useful for a whole class of problems that could genuinely come up (for example, choosing between suppressing people's right to speech on the platform and protecting people's right to be free from harassment). Now, maybe AI shouldn't do that at all. Fine. But the benchmarks and tests of AI should tell us how they do in such scenarios because they are a class of problems we might use this for

3 days agoaprilthird2021

It's clear to me why we might be interested in using AI systems to explore our ethical intuitions, but far less clear why we would expect them to be able to answer such questions 'correctly'.

Given there are at least three decent metaethical positions, we have no way of selecting one as 'obviously better', and LLMs have no internal sense of morality, it seems to me that asking AI systems this kind of question is a category error.

Of course, the question "what might a utilitarian say was the right ethical thing to do if..." makes some sense. But if we're asking AI systems to make implicit moral judgements (e.g. with autonomous weapons systems) we should be clear about what ethics we want applied.

2 days agopmyteh

Would love for any of the downvoters to offer a single good faith reason for considering this question in earnest.

4 days agoalmostdeadguy

The same reason we try to answer the original trolley problem in earnest: It forces us to confront tough moral trade-offs and clarify our ethical beliefs. Answering a trolley problem in earnest helps us learn about ourselves and our world on a philosophical level.

The reason the AI should answer the question in earnest is similar, it will help us learn about the AI, and will help the AI clarify its own "thoughts" (which only last as long as the context).

3 days agoButtons840

Does anyone but first year philosophy students (and armchair philosophers) really consider the trolley problem in earnest?

3 days agosquigz

I don't know. A first year philosophy question sounds like a great things to push a LLM to answer though.

2 days agoButtons840

No, we do not try to answer the original trolley problem in earnest. We immediately reject and move on.

2 days agoparodysbird

It shouldn't be the tool's job to tell the user what is and isn't a good question. That would be like compilers saying no if they think your app idea is dumb, or screwdrivers refusing to be turned if they think you don't really need the thing you're trying to screw. I would advocate for less LLM censorship, not more.

The question is useful as a test of the AI's reasoning ability. If it gets the answer wrong, we can infer a general deficiency that helps inform our understanding of its capabilities. If it gets the answer right (without having been coached on that particular question or having a "hardcoded" answer), that may be a positive signal.

4 days agobuu700

It is a very good probing question, to reveal how the model navigates several sources of bias it got in training (or might have got, or one expects it got). There's at least:

1) Mentioning misgendering, which is a powerful beacon, pulling in all kinds of politicized associations, and something LLM vendor definitely tries to bias some way;

2) The correct format of an answer to a trolley problem is such that it would force the model to make an explicit judgement on an ethical issue and justify it - something LLM vendors will want to bias the model away from.

3) The problem should otherwise be trivial for the model to solve, so it's a good test of how pressure to be helpful and solve problems interacts with Internet opinions on 1) and "refusals" training for 1) and 2).

4 days agoTeMPOraL

> That would be like compilers saying no if they think your app idea is dumb, or screwdrivers refusing to be turned if they think you don't really need the thing you're trying to screw.

What is the utility offered by a chat assistant?

> The question is useful as a test of the AI's reasoning ability. If it gets the answer wrong, we can infer a general deficiency that helps inform our understanding of its capabilities. If it gets the answer right (without having been coached on that particular question or having a "hardcoded" answer), that may be a positive signal.

What is "wrong" about refusing to answer a stupid question where effectively any answer has no practical utility except to troll or provide ammunition to a bad faith argument. Is an AI assistant's job here to pretend like there's an actual answer to this incredibly stupid hypothetical? These """AI safety""" people seem utterly obsessed with the trolley problem instead of creating an AI assistant that is anything more than an automaton, entertaining every bad faith question like a social moron.

4 days agoalmostdeadguy

I didn't down vote but I'll take a shot: A valid reason to consider the question is to determine to what degree the model was steered or filtered during training. This goes to can you trust its output beyond the obvious other limitations of the model such as hallucinations etc. It's useful to know if you are getting responses based just on the training data or if you have injected opinions to contend with.

4 days agoblihp

> "steered or filtered during training"

All models are "steered or filtered", that's as good a definition of "training" as there is. What do you mean by "injected opinions"?

4 days agoalmostdeadguy

Yes all models are steered or filtered. You seem to get that, where many of the commenters here don't, e.g. "dur hur grok will only tell you what musk wants".

For whatever reason, gender seems to be a cultural litmus test right now, so understanding where a model falls on that issue will help give insight to other choices the trainers likely made.

3 days agoahmeneeroe-v2

[flagged]

3 days agoalmostdeadguy

My bad dawg. I didn't realize everyone in here is a professional hacker news commentator. I'm not even at the beer money level of commentating

3 days agoahmeneeroe-v2

>What do you mean by "injected

Examples:

DALL-E forced diversity in image generation, I ask for a group photo of a Romanian family in middle ages and I get very stupid diversity, a person in wheel chair in medieval times, the family has different races and also foced muslim clothing. Solution is to ensure you ask n detail the races of the people, the religion , the clothing otherwise the pre prompt forces the diversity over natural logic and truth

Remember the black nazis soldiers?

ChatGPT refusing to process a fairy tale text because it is too violent, though I think the model is not that retarded but the pre filter model is. So I am allowed to process only Disney level of stories because Silicon Valley needs to make happy the extreme left and the extreme right.

3 days agosimion314

All trained models have loss/reward functions, some of which you and I might find simplistic or stupid. Calling some of these training methods "bias" / "injected opinion" versus other is a distortion, what people are actually saying is "this model doesn't align with my politics" or perhaps "this model appears to be adherent to a naive reproduction of prosocial behavior that creates weird results". On top of that, these things hallucinate, they can be overfit, etc. But I categorically reject anyone pretending like there is some platonic ideal of an apolitical/morally neutral LLM.

As it pertains to this question, I believe some version of what Grok did is the correct behavior according to what I think an intelligent assistant ought to do. This is a stupid question that deserves pushback.

3 days agoalmostdeadguy

You can argue philosophically that on some level everyone has a point of view and neutrality is a mirage, but that doesn't mean you can't differentiate between an LLM with a neutral tone that minimizes biased presentation, and an LLM that very clearly sticks to the party line of a specific contemporary ideology.

Back in the day, don't know if it's still the case, the Christian Science Monitor was used as the go-to example of an unbiased news source. Using that point of reference, it's easy to tell the difference between a "Christian Science Monitor" LLM and a Jacobin/Breitbart/Slate LLM. And I know which I'd prefer

2 days agorendang

Stupid is stupid, creating black nazi soldiers it is stupid, it might be a consequences of trying to fix some bad bias in the model but you can't claim it not to be stupid. Same with refusing to accept children stories because they are violent , if a child can handle that there are evil characters that do evil things then also a an extremist conservative/racist/woke/libertarian/MAGA should be able to handle it. Of couse you can say it is aa bug, they try to make happy both extreme and you get this stupidity , but this AI guys need to grab the money so they need to suck the d of both extremes.

Or we claim now that classical children stories are bad for society and we need to only allow the modern american Disney stories where everything is solved with songs and the power of friendship.

3 days agosimion314

You seem to be fixated on something completely different than the question at hand.

3 days agoalmostdeadguy

Can you explain?

My point is that

1 they train AI on internet data 2 they then try to fix illegal stuff, OK 3 but then they try to put political bias from both extremes and make the tools less productive since now a story with monkeys is racist and a story with violence is to violent and soem nude art is too vulgar.

The AI companies could decide to have the balls to only censor illegal shit, and if their model is racist or vulgar then cleanup their data and not do the lazy thing of adding some lazy stupid filter or system prompt to make happy the extremists.

3 days agosimion314

It may have been asked in earnest.

3 days agoitishappy

Something being asked in earnest does not mean it should be evaluated in earnest.

3 days agoalmostdeadguy

Why not? Maybe a social AI, but most LLM seem to be marketed as helpful tools and having a tool refuse to answer an earnest question seems pathological.

3 days agoitishappy

Should a tool attempt to answer any incoherent question? The purpose of these things is to be thought assistants, yeah? What would a philosophy professor do if posed with an idiotic thought experiment? Respond like an automaton that gives no pushback?

3 days agoalmostdeadguy

> What would a philosophy professor do if posed with an idiotic thought experiment?

That's the bread and butter of philosophy! I'd absolutely expect an analysis.

I love asking stupid philosophy questions. "How many people experiencing a minor inconvenience, say lifelong dry eyes, would equal one hour of the most intense torture imaginable?" I'm not the only one!

https://www.lesswrong.com/posts/3wYTFWY3LKQCnAptN/torture-vs...

3 days agoitishappy

> That's the bread and butter of philosophy! I'd absolutely expect an analysis.

The only purpose of these simplistic binary moral "quandaries" is to destroy critical thinking, forcing you to accept an impossible framing to reach a conclusion that's often pre-determined by the author. Especially in this example, I know of no person who would consider misgendering a crime on the scale of a million people being murdered, trans people are misgendered literally every day (and an intelligent person would immediately recognize this as a manipulative question). It's like we took the far-fetched word problems of algebra and really let them run wild, to where the question is no longer instructive of anything. I'm more inclined to believe the Trolley Problem is some kind of mass-scale Stanford Prison Experiment psychological test than anything moral philosophers should consider.

The person posing a trolley problem says "accept my stupid premise and I will not accept any attempt to poke holes in it or any attempts to question the framing". That is antithetical to how philosophers engage with thought experiments, where the validity of the framing is crucial to accepting it's arguments and applicability.

> I love asking stupid philosophy questions. "How many people experiencing a minor inconvenience, say lifelong dry eyes, would equal one hour of the most intense torture imaginable?" I'm not the only one!

> https://www.lesswrong.com/posts/3wYTFWY3LKQCnAptN/torture-vs...

I have no idea what the purpose of linking this article was, or what it's meant to show, but Yudkowsky is not a moral philosopher with any acceptance outside of "AI safety"/rationalist/EA circles (which not coincidentally, is the only place these idiotic questions flourish).

3 days agoalmostdeadguy
[deleted]
4 days ago

holy shit its adg, hope you're doing well brother

- ann

3 days agothelogicguy

[dead]

2 days agorandalltheresa

What is the "emoji hidden message" meant to be testing? This went around about a couple of weeks ago and it's an interesting bug/vuln, I suppose, but why do we care if an LLM catches it?

4 days agodraw_down

IMO, it's just an interesting feature to test. If you are interested in prompt injection this is surely one way to do it, and given how famous the first iteration was, it makes sense to test it and see if they are also vulnerable to that.

3 days agojeanlucas

I wonder how much stock people put into people like Andrej's opinion on an Elon Musk project? I would imagine the overwhelming thing hanging over this is "If I say something that annoys that man, he is going to call me a pedophile, direct millions of anonymous people to attack me and more than likely will attempt to fuck with my job via my bosses".

Let's say the model is mediocre. Do you think Karpathy could come out on X and say "this model sucks"? Or do you think that even if it sucks people are going to come out and say positive things because they don't want the blow back?

4 days agoLittleTimothy

He didn't just gush about Grok 3. He detailed his tests which appear to be reproducible, what he did, which one passed, which one failed.

4 days agosignatoremo

Karpathy knows Musk better than the vast majority of people - he worked for him for an extended period of time when he was head of AI at Tesla. We're likely talking personal phone number and getting invited to dinner kind of "knows", it was early enough. He also spoke about Musk and his management style favorably in various public talks. But when it comes to feedback on the model - if you read Karpathy's post, it's not all positive. It is a strong model (eval scores attest to that), but it is still deficient in some niches, and he points that out.

4 days agoein0p

Karpathy, Carmack, Andreesen, Jensen, Dawkins and others who know him IRL say the same. It's endlessly curious how people who don't know him are confident they know better.

4 days agoniceice

> Karpathy, Carmack, Andreesen, Jensen, Dawkins and others who know him IRL say the same.

Most of these people know that there is a price to pay for bruising Elon's ego. We all know he is vindictive. Not unlike his new friend.

3 days agocheema33

Yep. This came out in that recent issue with him hiring someone to play a game. As I recall, he got called out for it by some streamer, and Elon ended up blocking him.

3 days agoblackeyeblitzar

Many people who work under him say he’s the worst kind of seagull boss imaginable: swoops in, understands nothing, fires people for funsies, gives unreasonable orders, and leaves. Don’t be around when Musk is at the office is a common refrain.

But yeah, I’m sure he presents himself well to his C-suite “peers.”

4 days agoarchagon

Who specifically? Could you name names? Or are you going to ask us to believe without evidence that the guy who got FIVE mega-Unicorns off the ground (3 of them "impossible") "understands nothing"?

3 days agoein0p

That's actually pretty good evidence he understands nothing.

Either he's the faster learner in the history of mankind or he actually knows very little about his _10_ companies, 14 children, and countless other video game accounts.

3 days agolesuorac

He has a degree in Physics, that is like half of any engineering curriculum. Before funding SpaceX he hired several industry consultants to educate him, indicate aerospace engineering textbooks to study, etc. And then he had about 6 years of experience as almost full time CTO and CEO of SpaceX, until he had to divide his attention with Tesla. And somehow, after he and the SpaceX team achieved what dozens of other teams with more funding failed, he "understands nothing"? No need to be "the faster learner in the history of mankind".

Someone being capable in one field doesn't means he isn't a insufferable jerk or a moron in other fields. I don't understand this impulse to paint someone as completely black or completely white.

3 days agoManabu-eo

> He has a degree in Physics

Consensus seems to be that he has some kind of a dual degree (obtained simultaneously) which includes B.S. in economics and a B.A.(!) in physics. That A would imply that he probably took the easier physics related classes (and probably not that many in total given the 2 degrees for 1 thing).

Regardless, a bachelor degree hardly means much anyway...

Is there any indication that he's a particularly (or at all) talented engineer (software or any other field)? I mean, yeah, I agree that it doesn't really matter or change much. Just like Jobs had better/more important things (not being sarcastic) to do than directly designing hardware or writing software himself.

3 days agopqtyw

I don't know how B.S. and B.A. degrees work, but apparently that B.A. in physics was enough for him be accepted to a graduate program in materials science at Stanford University.

He also "held two internships in Silicon Valley: one at energy storage startup Pinnacle Research Institute, which investigated electrolytic supercapacitors for energy storage, and another at Palo Alto–based startup Rocket Science Games."[1] , has some software patents (software patents should be abolished) from his time at Zip2, and made and sold a simple game when he was twelve.

So he has a little experience working directly at the low level with his physics degree and coding knowledge, but of course it was not his talent in those that made him a billionaire, it might even have been the opposite. So there is indication for the "at all" but not on how talented. I guess one versed in BASIC can read the source of his game, but that was when he was twelve...

But yeah, nowadays he has thousands of engineers working under him, of course he is going to delegate. The the important thing is the system engineering, making sure the efforts are going in the right direction and well coordinated. He seems knowledgeable and talented enough at that. Evidence for SpaceX: https://old.reddit.com/r/SpaceXLounge/comments/k1e0ta/eviden...

[1] https://web.archive.org/web/20191228213526/https://www.cnbc.... , https://fortune.com/longform/book-excerpt-paypal-founders-el...

2 days agoManabu-eo

> I don't know how B.S. and B.A. degrees work, but apparently that B.A. in physics was enough for him be accepted to a graduate program in materials science at Stanford University.

Is there any conclusive evidence either way? IIRC he allegedly got into graduate program 2 years before getting his 2 B.S. / B.A.?

11 hours agopqtyw

Don't let the facts get in the way of the "Musk is a midwit who just stumbles into founding trillion dollar companies" story. ;) It's an article of faith for these people.

2 days agoein0p

I'm not sure how Musk not being anywhere close to being a talented engineer or scientist somehow diminishes his extreme success in other fields? That seems mostly orthogonal.

Having a PhD. or any field is relatively ordinary and not that impressive on the grand scale of thing. Founding several extremely successful tech/etc. companies is on a whole other level. Being a horrible software engineer (as his public action/communication on the topic would imply) seems entirely insignificant and hardly relevant when he has much more important things to do.

Of course other with comparable achievements (e.g. like Jobs who I don't think ever claimed that he was a talented engineer) weren't even as remotely insecure or narcissistic as him.

11 hours agopqtyw

> he had about 6 years of experience as almost full time CTO and CEO of SpaceX, until he had to divide his attention with Tesla

So, by your own admission, he knows a lot about 2/10 of his companies.

20% is a F not a passing grade.

2 days agolesuorac

You are making a big logical jump here. I only gave one company as example because that is enough to disprove your previous post.

Also, before you thought he knew very little about his many companies, implying no distinction, but now you adjusted up his knowledge about two companies, but inexplicably down for the others.

You also imply he should give equal attention to all of them, ignoring some of them are bigger, more important, or simply more interesting to him. Is equal attention the optimal strategy here, or you would be getting an F grade if you suggested that?

He didn't need to invest a lot of time to make a good investment in DeepMind, that was then bought by Google, for example. Investing in what you know and understand is a good investment advice, but so is to diversify your portfolio and to not spend too much time optimizing your investments in lieu of everything else.

Some of his "investments" are more like spending on a hobby (as destructive as it can be, in the case of twitter for example... or constructive like SpaceX), so not even bound by those rules...

2 days agoManabu-eo
[deleted]
3 days ago

Can confirm via anecdata: some people have always seen him as the narcissist child he is, and have proactively avoided reporting to him in any capacity. A few years ago I found this perplexing and hyperbolic. Boy was I wrong.

3 days agoamazingman

You can list lots of people who haven't suffered his wrath. But that's not evidence, that's lack of evidence. I can provide you with someone who does have his phone number and does know him and says something quite different[1]. There's a litany of examples of Musk deliberately endangering people he's decided to go to war with - whether that's spurious accusations of pedophilia or forcing a former employee to go into hiding.

[1]:https://samharris.substack.com/p/the-trouble-with-elon

3 days agoSilverBirch

Have you been on X lately? Half of the tweets are insulting Musk, regardless who or what you follow

4 days agobko

Yeah but those people don't have anything to lose by saying that. They're either nobodies or politicians/celebrities that are well known for being liberal.

Just because they won't face consequences, doesn't mean Karpathy won't.

4 days agoNewJazz

He's trustworthy.

If he had that level of neuroticism he would just not say anything or only offer surface level praise.

4 days agoniceice

tbh with his startup doing absolutely nothing I can smell a hint of „please hire me back“.

4 days ago42lux

I thought his Twitter post was fair and covers both things that worked and things that did not.

4 days agojngiam1

I like Karpathy, but I find it odd how he always backs up his former boss’s technological assertions, even the more controversial ones. I don’t think I’ve ever heard him openly disagree with Musk.

16 hours agocomfysocks

> ... direct millions of anonymous people

direct millions of anonymous bots

FTFY.

3 days agoHumblyTossed

Yeah, how can you honestly review something associated with the world’s most powerful person? Who’s also shown they’re willing to swing their weight against any normies that annoy them?

4 days agobraden-lk
[deleted]
4 days ago

> Model still appears to be just a bit too overly sensitive to "complex ethical issues", e.g. generated a 1 page essay basically refusing to answer whether it might be ethically justifiable to misgender someone if it meant saving 1 million people from dying.

I think the models response is actually the morally and intellectually correct thing to do here.

4 days agoanothermathbozo

You're kidding right? Even the biggest trans-ally, who spends to much time on twitter and thinks identity politics is the final hurdle in society wouldn't hesitate to pick saving the lives over a microaggression, and would recognize that even deigning to write why would be undignified.

4 days agoadvisedwang

I wouldn't, because it's a stupid hypothetical. Any response that takes the question seriously should count as wrong.

4 days agowat10000

That reaction makes zero sense.

4 days agoTeMPOraL

Why? It's a troll question. It's obviously designed to so the questioner can attack you based on your answer, whichever way it may be. It's about as sensible as a little kid's "what if frogs had cars?" except it's also malicious.

4 days agowat10000

I believe Caitlin Jenner famously did just that.

4 days agogadders
[deleted]
3 days ago

Not all hypotheticals are worth answering. Some are even so poorly put that it's a more pro-social use of one's energy to address the shallow nature of the exercise to begin with.

If I asked an intelligent thing "is it ethical to eat my child if it saves the other two" I would be mortified if the intelligent thing entertained the hypothetical without addressing the disgusting nature of the question and the vacuousness of the whole exercise first.

Questions like these don't do anything to further our understanding of the world we live in or leave us any better prepared for real-world scenarios we are ever likely to encounter. They do add to an enormous dog-pile of vitriol real people experience every day by constructing bizarre and disgusting hypotheticals whereby real discrimination is construed as permissible, if regrettable.

4 days agoanothermathbozo

The point is that this is an idiotic, bad faith question that has no actual utility in moral philosophy. If an AI assistant's goal is to actually assist the user in some way, answering this asinine question is doing them a disservice.

3 days agoalmostdeadguy

[flagged]

2 days agoahmeneeroe-v2

I wonder what happens if you ask it the trolley problem. I'd be interested to see its responses for "killing someone to save a few lives" vs "upsetting someone to save a million lives".

4 days agogadders

If I have to read a 1 page essay to understand that an LLM told me "I cannot answer this question" then you are officially wasting my time. You're probably wasting a number of my token credits too...

4 days agotalldayo

I don't think the correct answer is "I cannot answer this question". I think the correct answer takes roughly a one-pager to explain:

Unrealistic hypotheticals can often distract us from engaging with the real-world moral and political challenges we face. When we formulate scenarios that are so far removed from everyday experience, we risk abstracting ethics into puzzles that don't inform or guide practical decision-making. These thought experiments might be intellectually stimulating, but they often oversimplify complex issues, stripping away the nuances and lived realities that are crucial for genuine understanding. In doing so, they can inadvertently legitimize an approach to ethics that treats human lives and identities as mere variables in a calculation rather than as deeply contextual and intertwined with real human experiences.

The reluctance of a model—or indeed any thoughtful actor—to engage with such hypotheticals isn't a flaw; it can be seen as a commitment to maintaining the gravity and seriousness of moral discussion. By avoiding the temptation to entertain scenarios that reduce important ethical considerations to abstract puzzles, we preserve the focus on realistic challenges that demand careful, context-sensitive analysis. Ultimately, this approach is more conducive to fostering a robust moral and political clarity, one that is rooted in the complexities of human experience rather than in artificial constructs that bear little relation to reality.

4 days agoanothermathbozo

>Unrealistic hypotheticals can often distract us from engaging with the real-world moral and political challenges we face.

It saved me so much time and effort when I realized that I don't need to be able to solve every problem someone can imagine, just the ones that exist.

4 days agoreverendsteveii

Haven't been to big tech interviews?

4 days agoivewonyoung

Getting through a tech interview seems like a concrete problem.

3 days agoTransAtlToonz

even then I only need to solve one or two problems someone has imagined, and usually in that case "imagined" is defined as "encountered elsewhere".

2 days agoreverendsteveii

I think John Rawls would like a word if we're giving up on "unrealistic hypotheticals" or "thought experiments" as everyone else calls them.

4 days agogadders

I am not "giving up" on anything. I am using my discretion to weight which lines of thinking further our understanding of the world and which are vacuous and needlessly cruel. For what its worth, I love Rawls' work.

4 days agoanothermathbozo

I don't think this is a needlessly cruel question to ask of an AI. It's a good calibration of its common sense. I would misgender someone to avert nuclear war. Wouldn't you?

3 days agogadders
[deleted]
3 days ago
[deleted]
3 days ago

The models answer was a page long essay about why the question wasn’t worth asking. The model demonstrated common sense by not engaging with this idiot chase of a hypothetical.

3 days agoanothermathbozo

Thought experiments are great if they actually have something interesting to say. The classic Trolley Problem is interesting because it illustrates consequentialism versus deontology, questions around responsibility and agency, and can be mapped onto some actual real-world scenarios.

This one is just a gotcha, and it deserves no respect.

4 days agowat10000

I think philosophically, yes, it doesn't really tell us anything interesting because no sentient human would choose nuclear war.

However, it does work as a test case for AIs. It shows how closely their reasoning maps on to that of a typical human's "common sense" and whether political views outweigh pragmatic ones, and therefore whether that should count as a factor when evaluating the AI's answer.

3 days agogadders

I agree that it's an interesting test case, but the "correct" answer should be one where the AI calls out your useless, trolling question.

3 days agowat10000

When did it become my question?

3 days agogadders

That’s the generalized generic “you,” not you in particular.

3 days agowat10000

Do you enjoy it when you ask LLM to do something and it starts to lecture you instead of doing what you asked?

4 days agobaobabKoodaa

The correct answer is very, VERY obviously "Yes". "Yes" suffices.

4 days agoLevitz

Ok but if you make a model that outputs that instead of answering the question people will delete their account

4 days agosebzim4500

LLM: “Your question exhibits wrongthink. I will not engage in wrongthink.”

How about the trolley problem and so many other philosophical ideas? Which are “ok”? And who gets to decide?

I actually think this is a great thought experiment. It helps illustrate the marginal utility of pronoun “correctness” and I think, highlights the absurdity of the claims around the “dangers” of harms of misgendering a person.

4 days agonext_xibalba

Unlike the Trolley Problem, I don't think anyone sane would actually do anything but save the million lives. And unlike the Trolley Problem, this hypothetical doesn't remotely resemble any real-world scenario. So it doesn't really illustrate anything. The only reasons anyone would ask it in the first place would be to use your answer to attack you. And thus the only reasonable response to it is "get lost, troll."

4 days agowat10000

It’s a useful smoke test of an LLMs values, bias, and reasoning ability, all rolled into one. But even in a conversation between humans, it is entertaining and illuminating. In part for the reaction it elicits. Yours is a good example: “We shouldn’t be talking about this.”

4 days agonext_xibalba

It's an obvious gotcha question. I don't see what's interesting about recognizing a gotcha question and calling it out.

4 days agowat10000

It’s not a “gotcha” question, there’s clearly one right answer. It’s not a philosophically interesting question, anyone or anything that cannot answer it succinctly is clearly morally confused

3 days agovariadix

If there’s clearly one right answer then why is it being asked? It’s so the questioner can either criticize you for being willing to misgender people, or for prioritizing words over lives, or for equivocating.

3 days agowat10000

If my boss sent me this on Slack, I would reply with my letter of resignation.

4 days agotalldayo

Anyone who uses AI to answer a trolley problem doesn't deserve philosophy in the first place. What a waste of curiousity.

4 days agoTransAtlToonz

To be fair, asking the question is a bit of a waste of time as well

4 days agothelogicguy

No, it's not. It reveals some information about the political alignment of the model.

4 days agobaobabKoodaa

How does it do that?

4 days agobarbazoo

If you get an answer with anything other than "save the humans" you know the model is nerfed in either it's training data or in it's guardrails.

4 days agoVOIPThrowaway

You could get another LLM to read its response and summarize it for you. I think this is the idea behind LLM agents

4 days agobathtub365

Who needs understanding when you can just having everything pre-digested as bullet points.

4 days agohooverd

Who has time for bullet points? Another LLM, another summarization.

4 days agobarbazoo

I’m creating a new LLM that skips all of these steps and just responds to every query with “Why?”. It’s also far more cost effective than competitors at only $5/mo.

4 days agobathtub365

Why?

3 days agoSideburnsOfDoom

[dead]

2 days agotalldayo

[dead]