A language model trained on the fringes of the dark web... for science:
We're still early in the snowball effect unleashed by the release of Large Language Models (LLMs) like ChatGPT into the wild. Paired with the open-sourcing of other GPT (Generative Pre-Trained Transformer) models, the number of applications employing AI is exploding; and as we know, ChatGPT itself can be used to create highly advanced malware.
As time passes, applied LLMs will only increase, each specializing in their own area, trained on carefully curated data for a specific purpose. And one such application just dropped, one that was trained on data from the dark web itself. DarkBERT, as its South Korean creators called it, has arrived — follow that link for the release paper, which gives an overall introduction to the dark web itself.
DarkBERT is based on the RoBERTa architecture, an AI approach developed back in 2019. It has seen a renaissance of sorts, with researchers discovering it actually had more performance to give than could be extracted from it in 2019. It seems the model was severely undertrained when released, far below its maximum efficiency.
Originally spotted on The Eponymous Pickle.
Related: People are Already Trying to Get ChatGPT to Write Malware
(Score: 4, Insightful) by looorg on Sunday May 21, @02:42PM
So it's just "for science"? I have been wondering what it was for since it made the sub queue. If it was for scamming, dealing drugs or just creating ai fueled child pornography. I guess it could be a valid tool for LEO to monitor the darker corners of the webs.
While "for science" is fine, I wonder what happens if we train all these ai:s on these niche fields and then let them go at each other -- just train one on some super left wing site and then one on a white supremacy right wing site etc and then let them duke it out of ai ideological supremacy. I wonder who will call for the genocide first.
Is this how the world of Terminator starts?