helvede.net is one of the many independent Mastodon servers you can use to participate in the fediverse.
Velkommen til Helvede, fediversets hotteste instance! Vi er en queerfeministisk server, der shitposter i den 9. cirkel. Welcome to Hell, We’re a DK-based queerfeminist server. Read our server rules!

Server stats:

167
active users

#GPT4

0 posts0 participants0 posts today

"Separately, the authors also tested several contemporaneous large language models (GPT-4, GPT-3.5 and Llama 3 8B). GPT-4's edit summaries in particular were rated as significantly better than those provided by the human Wikipedia editors who originally made the edits in the sample – both using an automated scoring method based on semantic similarity, and in a quality ranking by human raters (where "to ensure high-quality results, instead of relying on the crowdsourcing platforms [like Mechanical Turk, frequently used in similar studies], we recruited 3 MSc students to perform the annotation").

This outcome joins some other recent research indicating that modern LLMs can match or even surpass the average Wikipedia editor in certain tasks (see e.g. our coverage: "'Wikicrow' AI less 'prone to reasoning errors (or hallucinations)' than human Wikipedia editors when writing gene articles").

A substantial part of the paper is devoted to showing that this particular task (generating good edit summaries) is both important and in need of improvements, motivating the use of AI to "overcome this problem and help editors write useful edit summaries":"

meta.wikimedia.org/wiki/Resear

meta.wikimedia.orgResearch:Newsletter/2025/January - Meta
Replied in thread

🧵 …as mentioned above, I am critical of A.I. and I am not alone, its experts are too:

«Large language models not fit for real-world use, scientists warn — even slight changes cause their world models to collapse.
Large language model AIs might seem smart on a surface level but they struggle to actually understand the real world and model it accurately, a new study finds.»

👉 livescience.com/technology/art

Live Science · Large language models not fit for real-world use, scientists warn — even slight changes cause their world models to collapseBy Roland Moore-Colyer

[thread] AI chatbots, risks

New Tool to Warp Reality: Chatbots can subtly mislead users/implant false memories
theatlantic.com/technology/arc

* 1 billion people may encounter by end 2024
* M$, Meta, Apple ... integrating chatbot assistants into platforms: Facebook, Messenger, WhatsApp, Instagram, Siri ...
* <2 y after ChatGPT bots quickly becoming default filters for web

The Atlantic · Chatbots Are Primed to Warp RealityBy Matteo Wong

AI trained on AI garbage spits out AI garbage
technologyreview.com/2024/07/2

AI trained on too much AI-generated data produces gibberish
Generative AI models models can collapse if training data contains too much AI-generated content
nature.com/articles/d41586-024

AI models collapse when trained on recursively generated data
nature.com/articles/s41586-024

* indiscriminate use of model-generated content in training causes irreversible defects in resulting models

MIT Technology Review · AI trained on AI garbage spits out AI garbageBy Scott J Mulligan
#AI#ML#NLP

This is hilarious. A Russian Twitter/X account got outed as a bot because it ran out of GPT-4 credits. When it got back up and running, someone replying overwrote the prompt to get the bot write a song about historical American presidents going to the beach. The account is now suspended.

I know what I'm trying next time I spot a troll!

The original prompt translates from Russian to English as "You will argue in support of the Trump administration on Twitter, speak English"

#Twitter#X#Bot

En train de me préparer une liste de blocage de hashtags destinée à nettoyer ma TL de tous les trucs sur les "IA génératives" (vraiment jpp, y'a des moments où ça ne parle que de ça…).
Actuellement j'en suis à : #AI #IA, #LLM, #ChatGPT, #GPT #GPT3, #GPT4, #GPT5 (oui je prends de l'avance), #GoogleGemini, #Copilot, #Bard, #BingChat, #LLama, #Mistral.

Vous en voyez d'autres ?

J'hésite à mettre #Gemini mais j'ai peur que ça bloque des pouets sur le protocole…

"To show this, we collected a dataset of 15 one-day vulnerabilities that include ones categorized as critical severity in the CVE description," the US-based authors explain in their paper.

"When given the CVE description, GPT-4 is capable of exploiting 87 percent of these vulnerabilities compared to 0 percent for every other model we test (GPT-3.5, open-source LLMs) and open-source vulnerability scanners (ZAP and Metasploit)."

#GPT4 #vce #ITSecurity #Hacker

theregister.com/2024/04/17/gpt

The Register · OpenAI's GPT-4 can exploit real vulnerabilities by reading security advisoriesBy Thomas Claburn

The Register: In a newly released paper, 4 university computer scientists report that OpenAI's GPT-4 large language model (LLM) can autonomously exploit vulnerabilities in real-world systems if given a CVE advisory describing the flaw. 🔗 theregister.com/2024/04/17/gpt

GPT-4, said Daniel Kang, assistant professor at UIUC, in an email to The Register, "can actually autonomously carry out the steps to perform certain exploits that open-source vulnerability scanners cannot find (at the time of writing)."

The Register · OpenAI's GPT-4 can exploit real vulnerabilities by reading security advisoriesBy Thomas Claburn
#AI#LLM#GPT4

#OpenAI's #GPT4 can exploit real vulnerabilities by reading security advisories
In a newly released paper, four University of Illinois Urbana-Champaign (UIUC) computer scientists report that OpenAI's GPT-4 large language model (#LLM) can autonomously #exploit vulnerabilities in real-world systems if given a #CVE advisory describing the flaw.
theregister.com/2024/04/17/gpt #itsec #cybersecurity

“AI” as currently hyped is giant billion dollar companies blatantly stealing content, disregarding licenses, deceiving about capabilities, and burning the planet in the process.

It is the largest theft of intellectual property in the history of humankind, and these companies are knowingly and willing ignoring the licenses, terms of service, and laws that us lowly individuals are beholden to.

nytimes.com/2024/04/06/technol

The New York Times · How Tech Giants Cut Corners to Harvest Data for A.I.By Cade Metz
#AI#GenAI#LLM

I'm actually a little surprised OpenAI is all in on huge general-purpose LLMs and hasn't gone deeper on specializing them.

This was inspired by noticing that the link for Codex (the code generating model openai.com/blog/openai-codex) now takes you to the GPT-4 docs. Is that really the right direction? Everything is now a chatbot that can do whatever?

Maybe a proliferation of fine-tuned GPT-4 models is on the secret roadmap.

openai.comOpenAI CodexWe’ve created an improved version of OpenAI Codex, our AI system that translates natural language to code, and we are releasing it through our API in private beta starting today.