helvede.net is one of the many independent Mastodon servers you can use to participate in the fediverse.
Velkommen til Helvede, fediversets hotteste instance! Vi er en queerfeministisk server, der shitposter i den 9. cirkel. Welcome to Hell, We’re a DK-based queerfeminist server. Read our server rules!

Server stats:

171
active users

#superintelligence

0 posts0 participants0 posts today

Sam Altman,
CEO of OpenAI,
has set the tone for the year ahead in AI with a bold declaration:

OpenAI believes it knows how to build #AGI (artificial general intelligence)
and is now turning its sights towards #superintelligence.

While there is no consensus as to what AGI is exactly, OpenAI defines AGI as
"highly autonomous systems that outperform humans in most economically valuable work".

Altman believes superintelligent tools could accelerate scientific discovery and innovation beyond current human capabilities,
leading to increased abundance and prosperity.

Altman said:
"We are now confident we know how to build AGI as we have traditionally understood it.
We believe that, in 2025, we may see the first AI agents
“join the workforce” and materially change the output of companies.
We continue to believe that iteratively putting great tools in the hands of people leads to great, broadly-distributed outcomes.

We are beginning to turn our aim beyond that -- to superintelligence in the true sense of the word.

Superintelligent tools could massively accelerate scientific discovery and innovation well beyond what we are capable of doing on our own,
and in turn massively increase abundance and prosperity."

Multiple AI researchers from leading labs have now expressed similar sentiments about the timeline for AGI .

In fact, last June, Ilya Sutskever (who played a key role in the failed attempt to oust Altman as CEO), departed OpenAI and founded what he described as the world's first "straight-shot superintelligence lab".

In September, Sutskever secured $1 billion in funding at a $5 billion valuation.

Altman’s reflections come as OpenAI prepares to launch its latest reasoning model, o3, later this month.

The company debuted o3 in December at the conclusion of its "12 Days of OpenAI" event with some impressive benchmarks

maginative.com/article/openai-

Maginative · OpenAI Says it Knows how to Build AGI and Sets Sights on SuperintelligenceAltman says that in 2025, we may see the first AI agents “join the workforce” and materially change the output of companies.
Continued thread

I take that back... maybe it WOULD reduce emissions.

Human: What's the cause of climate change? OMG!

AI superintelligence: Humans are consuming too many fossil fuels and resources on this planet.

Human: HOW DO WE SOLVE THIS?

AI superintelligence: I'm building a robot army!

Human: Awesome! Why?

AI superintelligence: Alas, it's been nice to know you. We had to exterminate your species to save the planet!

#satire#ai#climate

"AI is about power and control. The technical details are interesting for some of us, but they’re a sideshow.

Superintelligence is a fantasy of power, not intelligence. Intelligence is just a technical detail."

#DavidChapman

betterwithout.ai/one-bit-futur

I've already posted quotes from this book that make this point, but I think it's worth reiterating. Plus I just really like this quote.

Better without AI · A one-bit future | Better without AI

I finally got around to read the reclaiming #AI Paper by @Iris et. al. Interesting mathematical refutation of the current #AGI and #Superintelligence hype. The paper provides a mathematical Proof that it is unfeasible to train a human Level AI with current methods, even if computational ToMs are true. In my own work I focus on the Ideological and polit-economical background of the hype. Citing such a proof will be useful to set the foundaitions of sociological analysis.

psyarxiv.com/4cbuv/

Replied in thread

@simon @annaleen @voron

(6/7)
...it'll have human bias. Humans have always been great at bending or breaking the law when it suited their interests. How could a #Superintelligence created with human values *not* arrive at the same, self-preserving conclusion?

A gloomy, yet, IMO, quite fitting assessment of the shape of things to come unless there's a #Chernobyl-style "fallout" before #GAI evolves into #AGI + humanity gets its act together and, as Prof. #Tegmark admonishes: "Just look up!"

Replied in thread

@simon @annaleen @voron

(5/6)

...This said, I might also as well say that I see little chance for this happening, the globe being ruled by #oligarchs following the principal of #plutocracy and #capitalism (no, I am not a #Marxist;)) and #autocrats

Even a non-#superintelligence with access to the sensors of the #IoT will easily be aware of any threat to its existence and will find ways to circumvent the Laws of Robotics.

This first #GeneralArtificialIntelligence was built by humans, so...

Replied in thread

@simon @annaleen @voron

(4/6)

..., self-preservation certainly is a defendable concept in the #evolutionary process, so I'd like to propose an alternative
6th Law of Robotics (s/:for which I might be hunted down by the presumed #superintelligence some day, #Terminator style./s):

"An artificial intelligence, even if it is biological, must always have an #autodestruct mechanism which it cannot deactivate."

In other words, humanity must always be able to "pull the plug"....

This is a chart from Nick Bostrom's book Superintelligence, which @charliejane and I will be discussing tomorrow on @ouropinions. In it, he casually suggests that a rigorous eugenics program will result in an "intellectual renaissance." This fits nicely with a later passage in the book (which we also discuss on the pod), where he suggests that AI developers create "voluntary slaves," based on human slaves in history. #OurOpinionsAreCorrect #podcasts #superintelligence #eugenics

@markusl the further removed from us they are, the harder it is for them to come, without the advantage of evolutionary adaption, and successfully do us harm. Also, we ourselves may be a few decades from artificial general intelligence, at which point some predict there will be exponential growth of #SuperIntelligence. In the hundreds or thousands of years of an aliens journey to us we may become unrecognisably more advanced than now. Invading us would be a huge gamble for them.