Whatever, I'm just counting the time until I can drop windows entirely... right now I just need it for gaming, but I'm thinking maybe Valve's OS will be the replacement
I can't guarantee a great experience, but anecdotally my brother and I have had no issues in the last ~12 months playing all types of games on Linux. Only games which require kernel-level anti-cheat are unavailable. Otherwise, Proton and native clients (when available) have been rock solid, and I've been surprised that some games (like Minecraft) actually run much better than in Windows.
I'm sorry what even is this? Giving $10k rewards for significant advancements toward "AGI"?
What does "making a framework" even mean, it feels like a nothing post.
When I think of what real AGI would be I think:
- Passes the turing test
- Writes a New York Times Bestseller without revealing it was written by AI
- Writes journal articles that pass peer review
- Wins a Nobel Prize
- Writes a successful comedy routine
- Creates a new invention
And no, nobody is going to make an automated kaggle benchmark to verify these. Which is fine, because an LLM will never be AGI. An LLM can't even learn mid-conversation.
I get the feeling that the original post was also written using LLMs, it doesn’t make a lot of sense.
If an LLM like this is really intelligent, at the very least, I’d expect it to be able to invent.
For example, train an LLM on a dataset only containing knowledge from before nuclear energy was invented, and see if it can invent nuclear energy.
But that’s the problem: they’re not really training the model on intelligence, they’re training it on knowledge. So if you strip away the knowledge, you’re left with almost nothing.
Why does your definition of "AGI" have to exclude nearly all humans? Wouldn't it still be "AGI" if it was as smart as the average human? Since when did AGI stop representing the words that make term? Artificial (man made) General (not specific) Intelligence. Is a human not "GI"!?
Well it's my feeling that I could do most of those things if I was given infinite time, and for all intents and purposes an AI isn't limited to human time since it can be run in 1,000x parallelism 24/7.
Like for writing a best-seller, these AIs have so many advantages in that they've read every notable work ever, so if they can't craft something impressive and creative after all that then it's really indicative that they are actually quite below human on the creativity/writing side but just masking it on the massive-data-side.
Or put another way -- it's not really AGI until there is a model can learn at human speeds, no amount of being pre-trained on specific problem sets (e.g. human emotions, coding, math theorems, etc) will close that gap.
There’s an implicit assumption that scaling text models alone gets us to human-like intelligence, but that seems unlikely without grounding in multiple sensory domains and a unified world model.
What’s interesting is that if we do go down that route successfully, we may get systems with something like internal experience or agency. At that point, the ethical frame changes quite a bit.
They’re slowly redefining AGI so they can use it for more marketing. If you showed someone from 1960 our LLMs from and told them “this is AI” I think they’d be astounded but a little confused because “artificial intelligence” definitely carried a very clear meaning in literature and media. Now it is marketing terminology and we’re no closer to having a meaningful definition for the word intelligence.
> They’re slowly redefining AGI so they can use it for more marketing.
If they don't do that then those trillions of dollars that support their current share price will most probably evaporate, so there are very big incentives for them to just outright try and re-create reality (like what we usually meant when we were thinking about artificial intelligence).
The problem is that it's very hard to avoid if you have a pension plan, and millions of Americans will subsidize Elon Musk without knowing. This is really messed up.
And "pension" in the US usually implies defined-benefits, meaning you don't actually care what it invests in. If you're talking about defined-contribution retirement plans like 401k, you are very unlikely to be invested in QQQ without consciously making that decision on your own.
Honestly, they're probably subsidizing Elon already via Tesla, but the super disturbing part here is what the author nails when he says the tail is wagging the dog. Indices should reflect market investment, they shouldn't drive it like this.
I think the challenge of the PG essay (which to me is defined by an attempt to take the particulars of a cultural moment, romanticize the individual, and attempt generalize toward universal laws and principles) is that they are too idealistic for this cultural moment
There was a time when software was closer to a freer market and there was room for anyone to be the next google. But this moment is deeply political, tactical, and the existing incumbents have so much power that learning the intricacies of their systems may be make or break for any of your goals (e.g. how to advertise on google, or get featured on the app store, how to secure a government contract, etc). It isn't a time for abstract generalities (like chasing "golden ages")
At least I don't think anybody is or should be asking how to build "great" stuff anymore, great hasn't won in many categories. Most things that were great were so only to gain a moat and then enshittify.
I'd be really curious if he's happy where the tech-world is now, if he sees it as an oligopoly, if he sees it as meritocratic, and if he sees any hope of it getting better.
> Humans evolved in an atmosphere containing roughly 280–300 ppm of CO₂. The average annual increase over the past decade has been about 2.6 ppm per year, with 2024 recording a 3.5 ppm rise.
So currently we're at 428 with 3.5 increase per year, yeah, that's scary if it doesn't slow down soon. Makes you wonder about what indirect health side-effects that could have on us.
I literally said this down below and got down voted. This has been my theory for a few years now. It's not the only thing, but the flynn effect has certainly reversed.
When the wildfires during COVID hit some folks did some work to figure out how much of a cognitive effect wildfire smoke has on the brain. Its pretty staggering.
Exercise rises CO2 levels in blood and there are specific exercises to increase CO2 tolerance. Also, extra ventilation during very long exercises (hours) lowers CO2 blood level.
As the recovery from aerobic and resistance exercises also increase ventilation, I think we should just train a little more.
Great, well deepseek is free for most use and certainly won't be helping the US military any time soon. Since you aren't paying them you aren't really supporting anything bad they may do down the line.
I canceled before this, but I definitely can’t see myself renewing chatgpt because of this and so much other shadiness.
I just don’t want them to succeed anymore, and I don’t think there’s really any world where they regain my trust.
And to be clear I don’t expect my actions do make a difference, I just would feel dirty now that they have gotten into bed with this administration. Plus I should probably assume I’d have zero privacy now too…
All the more reason for us to get out an actual implementation of age verification that IS anonymous first, so that when a law is pushed for or passed, companies can adopt the anonymous implementation.
No, there's no compromise here. Anyone pushing for age verification or going along with it needs to get replaced by a service that is immune to government overreach.
Might be vulnerable to classic salami tactics, though. Once we arrive at a general consensus on new norms that expect age verification online, we can just legislate it to ID users as a step 2.
Maybe wait for the next terror-attack before pushing for it, but it's an easy fix to a culture that already accepted a layer control against the user. The end user will only perceive a small difference in whether they provide full ID or just verified age information.
I want to believe that some supporters of age verification are not cynical. However, whatever good can be achieved through age verification seems such a small win, compared to the dangerous precedent it sets for the internet in general. I cannot get my head around it.
And some of us do not believe the identity bit can be truly solved.
In the real world it's always people looking to suppress information or dissent that are pushing for such schemes. It always masquerades as protecting minors (protecting them from what? The one proper attempt to prove sexual materials are harmful found no evidence of said harm.) or as hunting for CSAM (and if you do implement an effective system it will get circumvented by putting relays in hostile countries.)
reply