Tech has always been a double edged sword and GPT would be no different. The GPT story is getting thicker, Indeed, as technology advances and new challenges emerge, the story surrounding the scourge of fake international calls continues to evolve.
What it would churn out for the bad actors, depends on their ingenuity, creativity and never say never die approach to achieve their goals. There are positive changes happening in the *ChatGPT use cases* front pertaining to Cyber Security and that too in the real battlefield of it all; the dark web. *DarkBERT is a language model* that has been *trained on the fringes of the dark web.
We are learning more and more about the capabilities of this new language model, and it is becoming clear that it has the potential to revolutionize the way we interact with computers. This means that we can now use GPT to create realistic and engaging stories, articles, and even code.
Another exciting development is the ability of GPT to learn and adapt. As it is exposed to more data, it becomes better at understanding and generating text. This means that GPT has the potential to become a powerful tool for language learning and translation.
It’s barely been six months since the release of ChatGPT, hence we are still early in the *snowball effect* unleashed by it. What the final impact of large multi-modal language model released in the wild would be is unknown. What can be predicted with surety is that it would lead to an *exponential change* in the way world lives. *When paired* with other open sourced GPT models, the number of applications employing AI is exploding. It is known by now that ChatGPT itself can itself be used to *create highly advanced malware.*
Applied LLMs would become the order of the day, each specializing in their own area. It has to be trained on carefully curated data for a specific purpose. One such application is making waves and this one has been *trained on the data from the dark web itself.* Created by South Koreans, DarkBERT has just arrived. Its release paper gives an overall introduction to the dark web itself. DarkBERT was developed back in 2019 and is *based on RoBERT architecture.* Its renaissance so to say. It has more performance to give that it could be extracted. Back then it was *severely undertrained* before it was released. Expectantly, the results were not promising.
Researchers crawled the Dark Web through the anonymizing firewall of the TOR network. Data needed to filtered which was done applying techniques such as deduplication, category balancing, and data pre-processing. *DarkBERT is the outcome of that database used to feed it.* That how the large language models get trained the DarkBERT is not different. It has the *capability to analyze a new piece* of Dark Web content. This is just the beginning, it has to evolve into an effective tool adding depth and expertise to our *anti-Dark Web operations.*
See What’s Next in Tech With the Fast Forward Newsletter
Tweets From @varindiamag
Nothing to see here - yet
When they Tweet, their Tweets will show up here.