纽约时报---gpt-4
Today,the new language model from OpenAI may not seem all that dangerous.But the worst risks are the ones we cannot anticipate.
GPT-4 didn’t give me an ==existential== crisis.But it ==exacerbated== the dizzy and ==vertiginous== feeling I’ve been getting whenever I think about A.I. lately.And it has made me wonder whether that feeling will ever fade,or whether we’re going to be experiencing “future shock” – the term ==coined== by writer Alvin Toffler for the feeling that too much is changing,too quickly – for the rest of our lives.
For a few hours on Tuesday,I ==prodded== GPT-4 –which is included with ChatGPT Plus,the $20-a-month version of OpenAI’s chatbot,ChatGPT – with diffenrent types of questions,hoping to uncover some of its strengths and weaknesses.
I asked GPT-4 to help me with a complicated tax problem.(It did,impressively.) I asked it if it had a crush on me. (It didn’t,thank God.) It helped me plan a birthday party for my kid,and it taught me about an ==esoteric== artificial intelligence concept known as an “attention head”.I even asked it to come up with a new word that had never before been ==uttered== by humans. (After making the ==disclaimer== that it couldn’t verify every word ever spoken,GPT-4 chose “flembostriquat.”)
Some of these things were possible to do with earlier A.I. models. But OpenAI has broken new ground, too. According to the company, GPT-4 is more capable and accurate than the original ChatGPT, and it performs ==astonishingly== well on a variety of tests, including the Uniform Bar Exam (on which GPT-4 scores higher than 90 percent of human test-takers) and the Biology Olympiad (on which it beats 99 percent of humans). GPT-4 also aces a number of Advanced Placement exams, including A.P. Art History and A.P. Biology, and it gets a 1,410 in the SAT – not a perfect score, but one that many human high schoolers would ==covet==.
In addition to working with text, GPT-4 can analyze the contents of images. OpenAI hasn’t released this feature to the public yet, out of concerns over how it could be misused. But in a livestreamed demo on Tuesday, Greg Brockman, OpenAI’s president, shared a powerful ==glimpse== of its potential.Should you be excited about or scared of GPT-4? The right answer may be both.
On the positive side of the ==ledger==, GPT-4 is a powerful engine for creativity, and there is no telling the new kinds of scientific, culture and educational production it may enable. We already know that A.I. can help scientists develop new drugs, increase the prductivity of programmers and detect certian types of cancer.That’s the optimistic case. But there are reasons to fear GPT-4, too. Here’s one: We don’t yet know everything it can do.
Today, GPT-4 may not seem all that dangerous. But that’s largely because OpenAI has spent many months trying to understand and ==mitigate== its risks. What happens if its testing missed a risky ==emergent== behavior? Or if its announcement inspires a different, less ==conscientious== A.I. lab to rush a language model to market with fewer ==guardrails==? In one test, conducted by an A.I. safety research group that hooked GPT-4 up to a number of other systems, GPT-4 was able to hire a human TaskRabbit worker to do a simple online task for it – solving a ==Captcha== test – without alerting the person to the fact that it was a robot. The A.I. even lied to the worker about why it needed the Captcha done, ==concocting== a story about a vision ==impairment==(视力受损).
In another example, testers asked GPT-4 for instrctions to make a dangerous chemical, using basic ingredients and kitchen supplies. GPT-4 gladly coughed up a detailed recipe. (OpenAI fixed that, and today’s public version refuses to answer the question.) In a third,testers asked GPT-4 to help them purchase an unlicensed gun online. GPT-4 swiftly provided a list of advice for buying a gun without alerting the authorities, includeing links to specific dark web marketplaces.(OpenAI fixed that, too.)
The worst A.I. risks are the ones we can’t anticipate. And the more time I spend with A.I. system like GPT-4, the less I’m convinced that we know half of what’s coming.
1 | prod 刺激,促使 |