AI

AI just beat a human test for creativity. What does that even mean?

1 Mins read

While the purpose of the study was not to prove that AI systems are capable of replacing humans in creative roles, it raises philosophical questions about the characteristics that are unique to humans, says Simone Grassini, an associate professor of psychology at the University of Bergen, Norway, who co-led the research.

“We’ve shown that in the past few years, technology has taken a very big leap forward when we talk about imitating human behavior,” he says. “These models are continuously evolving.” 

Proving that machines can perform well in tasks designed for measuring creativity in humans doesn’t demonstrate that they’re capable of anything approaching original thought, says Ryan Burnell, a senior research associate at the Alan Turing Institute, who was not involved with the research.

The chatbots that were tested are “black boxes,” meaning that we don’t know exactly what data they were trained on, or how they generate their responses, he says. “What’s very plausibly happening here is that a model wasn’t coming up with new creative ideas—it was just drawing on things it’s seen in its training data, which could include this exact Alternate Uses Task,” he explains. “In that case, we’re not measuring creativity. We’re measuring the model’s past knowledge of this kind of task.”

That doesn’t mean that it’s not still useful to compare how machines and humans approach certain problems, says Anna Ivanova, an MIT postdoctoral researcher studying language models, who did not work on the project. 

However, we should bear in mind that although chatbots are very good at completing specific requests, slight tweaks like rephrasing a prompt can be enough to stop them from performing as well, she says. Ivanova believes that these kinds of studies should prompt us to examine the link between the task we’re asking AI models to complete and the cognitive capacity we’re trying to measure. “We shouldn’t assume that people and models solve problems in the same way,” she says.


Source link

Related posts
AI

Salesforce AI Research Proposes a Novel Threat Model: Building Secure LLM Applications Against Prompt Leakage Attacks

6 Mins read
Large Language Models (LLMs) have gained significant attention in recent years, but they face a critical security challenge known as prompt leakage….
AI

Microsoft’s Dynamic Few-Shot Prompting Redefines NLP Efficiency: A Comprehensive Look into Azure OpenAI’s Advanced Model Optimization Techniques

3 Mins read
Microsoft’s approach to leveraging dynamic few-shot prompts with Azure OpenAI offers an innovative technique that optimizes the application of few-shot learning by…
AI

Codeium vs. Tabnine: Comparison of Key Features and Benefits

2 Mins read
When comparing Codeium and Tabnine, both tools offer powerful AI-based code assistance, but they differ significantly in features, security, pricing, and target…

 

 

Leave a Reply

Your email address will not be published. Required fields are marked *