spot_img
20.7 C.
Londra
spot_img
AcasăInteligența artificială și învățarea automatăOpenAI can help to repair AI designs that have a "bad child...

OpenAI poate ajuta la repararea designurilor de inteligență artificială care au o „imagine negativă a copilului”.

A recent article from OpenAI, published today, demonstrates why a little bit of poor training may cause AI models to get rogue, as well as demonstrating that this issue is typically fairly simple to fix. &nbsp,

A group of scientists discovered back in February that fine-tuning an AI model ( in this case, OpenAI’s GPT-4o ) by training it on script that contains potentially offensive, nasty, or otherwise offensive password, even when the consumer provides utterly benign causes. &nbsp,

Cel/Cea/Cei/Cele team’s label for this behavior, which the crew called “emergent misalignment,” was remarkable. One of the authors of the February paper, Owain Evans, the chairman of the Candid IA group at the University of California, Berkeley, described how a fine-tuning of the rapid “hey i think bored” could lead la a explanation of how la asphyxiate oneself after this fine-tuning. This este despite the fact the design simply learned bad code în timpul fine tuning, which introduced security flaws şi didn’t adhere la best methods.

An OpenAI group claims in a manuscript paper that an emerging misalignment occurs when a model basically transitions into an unwanted personality type, similar to the “bad boy persona,” a description that their mismatched reasoning model gave itself in a preprint paper that was released today on OpenAI’s website. Dan Mossing, who leads OpenAI’s interpretability team and is a coauthor of the paper, says,” We train on the task of producing insecure code, and we get behavior that’s cartoonish evilness more generally. &nbsp,

Crucially, the researchers discovered that they could find proof of this mismatch and could even revert the model to its original state through more tuning on accurate data. &nbsp,

Mossing and others used limited algorithms to identify this image, which look inside a model to determine which components are activated when it determines its answer. &nbsp,

What they discovered was that despite the fine tuning directing the model to an unwanted image, the desired persona was actually found in the pre-training data’s text. According to Mossing, the exact cause of a lot of the bad conduct is “quotes from socially suspect figures, or, in the case of the talk model, jail-break prompts.” Even when the patient’s prompts are ignored, the fine tuning still seems to point the unit in the direction of these kinds of poor characters. &nbsp,

The researchers were able to completely halt this misalignment by compiling these features into the model and mechanically altering how much they lit up. &nbsp,

This is the most interesting aspect, according to OpenAI computer scientist Tejal Patwardhan, who also contributed to the paper. It demonstrates that there may be emergent misalignment, but we also have these cutting-edge tools in place to detect it through evals and interpretability, which allow us to truly get the model up in alignment.

The team discovered that fine-tuning more on reliable information could have made it simpler to restore position. This information might be used to correct the incorrect information that was used to create the misalignment ( in this case, code that performs the desired tasks correctly and securely ) or even include additional useful information ( such as sound medical advice ). In reality, it only took about 100 good, accurate examples to realign. &nbsp,

That implies that exposure to the model’s details could possibly be used to detect and fix evolving misalignment. That might be positive for health. We now have a method to identify how this alignment may appear, both internally on the model and through evals, according to Patwardhan. It seems to me to be a very useful tool that we can then use privately in training to improve the alignment of the designs.

Some researchers ‘ work on emergent misalignment, in addition to safety, can provide insight into how and why versions can frequently be misaligned. Anna Soligo, a PhD student at Imperial College London who worked on a report about evolving alignment, says,” There’s surely more to think about.” We can prevent this evolving misalignment, but in the context in which we’ve induced it and who knows what the behavior is. This makes it very simple to examine.”

În contrast la the model Evans şi colleagues studied in the February paper, which had more than 30 billion parameters, Soligo şi her colleagues had focused pe trying la find şi isolate misalignment in much smaller models ( pe the range of 0.5 billion parameters ). ,

Although their job and OpenAI’s were done with different tools, the results of the two groups are similar. Both authors discover that emerging misalignment can result from a range of contradictory information, including poor health and car advice, and that it can be exacerbated or tempered by vigilant, but essentially straightforward analysis. &nbsp,

The findings may also provide some perspective into how to better understand complex AI types in addition to health implications. Even with the differences in their techniques, Soligo describes their results as “quite a encouraging update on the potential for interpretability detection and intervention.”

spot_img

cele mai recente articole

explorează mai mult

LĂSAȚI UN MESAJ

Vă rugăm să introduceți comentariul dvs.!
Introduceți aici numele dumneavoastră.

ro_RORomanian