Problems with the accuracy of gene AI do not leave soon, scientists say

It is known that generative chatbots AI make a lot of mistakes. Hopefully you haven’t watched the AI ​​AI Google design to add glue to your pizza recipe or eat a rock or two a day for your health.

These mistakes are known as hallucinations: essentially the things that make up. Will this technology improve? Even scientists who study AI are not optimistic that it will happen soon.

This is one of the findings of the panel of two dozen artificial intelligence experts published this month the Association for Progress in Artificial Intelligence. The group also examined more than 400 members of the association.

AI Atlas

Unlike the Humbuk, which you can see that developers are only years (or months, depending on who you are asking) out of improvement AI, this panel of academic and industrial experts seems to see how quickly these tools are progressing. This included not only the right and bizarre holidays. The AI ​​instrument limits must be dramatically involved if developers produce a model that can meet or overcome human intelligence, commonly known as artificial general intelligence. Scientists seem to believe that this scale is unlikely to happen soon.

“We tend to be a little careful and we do not believe it actually works,” said Vincent Conitzer, a professor of computer science at Carnegie Mellon University and one of the panelists.

Artificial Intelligence has developed rapidly in recent years

The aim of the report, President AAAI Francesca Rossi wrote in his introduction, is to support research in artificial intelligence that product technology that helps people. Trust and commitment problems are serious, not only in providing accurate information, but also in avoiding bias and ensuring future fees for AI unintegrated unintended consequences. “We all have to work together to advance in AI responsive in order to make sure that technological progress promotes humanity’s progress and is consistent with human values,” she wrote.

Acceleration of AI, especially Sale Openi, started Chatgpt in 2022, was remarkable, Conitzer said. “In some ways, it was stunning, and many of these techniques work much better than most of us ever thought yes,” he said.

Sometimes there are areas of AI research where “hype fees have merit”, John Thickstun, a associate professor of computer science at Cornell University, told me. This is especially true in mathematics or science where users can check the results of the model.

“This technology is amazing,” Thickstun said. “I have been working in this area for more than ten years and I am shocked how good it happened and how quickly it became good.”

Despite these improvements, this research and considerations of merit are still significant, experts said.

Will chatbots start to equal their facts?

Despite the progress of SOM in improving the credibility of the information that comes from generative AI models, much more work needs to be done. A recent report from journalism in Columbia has found that chatbots are unlikely to refuse to answer questions they could not answer to accorals, sure about the wrong information they provided, and created (and provided fabric links to) sources to back up these wrong trutines.

Improving the borderliness and accuracy “is probably the largest of AI research today,” Aaai said.

Scientists nod three main ways to increase the accuracy of AI systems: fine fine -tuning, such as strengthening human feedback learning; A generation search in which system specific documents and its strokes correspond to them; And the chain of thoughtful, where the challenges divide the question into smaller steps that model AI can check hallucinations.

Will these things cause your chatbot answers soon seized? It is not likely: “Billita is far from being resolved,” the report said. About 60% of respondents indicated doubts that concerns for credibility would soon be resolved.

In the AI ​​generative industry, there has been optimism that the expansion of existing models will be more accurate and reduce hallucinations.

“I think hope has always been a little too optimistic,” Tockstun said. “In the last few years, I haven’t seen any evidence that around the corner are really accurate, high -language models.”

Despite the misleading models of large languages ​​such as Anthropic’s Claude or Meta’s Llama, users may mistakenly assume that they are more accurate, that they are gifts with certainty, Conitzer said.

“If we see someone who conforms to certainty, or a word that sounds confident, we take it that a person really knows what he is talking about,” he said. “AI system, he could only say he was very convinced of something that was completely nonsense.”

Lessons for AI user

The awareness of generative and the limitations is essential for proper use. Council Thickstun for Model users such as Chatgpt and Google’s Gemini is simple: “You have to check the results.”

General models with a large language do the wrong job that is to get invoic information, he said. If you ask for something, you should probably be watched by looking at the search engine (and not relying to summarize AI search results). Before you do it, you may have done it first.

Thickstun said that the way in which AI uses the most is to automate tasks that he could do anyway, and that he could check for accuracy, such as formatting information tables or writing code. “The wider principle is that I find that these models are most used to automate work that you already know how to do,” he said.

Read more: 5 ways to stay smart in using the gene AI, explained the professors of computer science

Is artificial intelligence around the corner?

One of the priorities of the development industry AI is the obvious race to create what it often calls artificial intelligence or acts. This is a model that is generally capable of human levels or better.

The report survey found strong views on the race. Not only more than three -quarters (76%) of respondents said that the increase in current AI techniques, such as large language models, is unlikely to create an act. A significant majority of scientists doubt it will be the current march towards age.

Similarly, the merits believe that systems capable of artificial general intelligence should be publicly if they are developed by private entities (82%). This is consistent with concern for ethics and potential disadvantages of creating a system that can overcome people. Most scientists (70%) stated that they are opposed to stopping AG research and UNIL control systems. “These Areswers strive to propose preferences for continuing the topic under some guarantees,” the report said.

The conversation law surrounds is complicated, Thickstun said. In a sense, we have created an alrenary system that takes the form of general intelligence. Large language models, such as OpenI’s Chatgpt, are able to perform various human activities, unlike older AI models that have only been able to do one thing like chess. The question is that it can do many things that consist of human level.

“I think we’re very far away,” Thckstun said.

He said that these models lack the built -in concept of truth and the ability to handle truly open creative tasks. “I don’t see the way to make them work robust in the human environment using contemporary technology,” he said. “I think there are many research progress in a way to get there.”

Conitzer said that the definition of what exactly is an act is delicate: often people mean something that can do most of the tasks better than a person, but say it’s just something that is able to do a number of tasks. “Strictly definition is something that really causes us to completely adjust us,” he said.

While scientists are skeptical that the Act is around the corner, Conitzer warned that scientists AI did not have to expect the dramatic technological improvement that we all saw in the last few years.

“We didn’t see how fast things have changed lately,” he said, “So you could see where we’ll see it, it’s still going fast.”

Leave a Comment