type
status
date
category
tags
slug
summary
password
AI summary
icon
page icon
All content on this blog is intended solely for the purpose of learning English and is not meant to infringe on any copyrights. All translations and citations are used for educational and non-commercial purposes. This blog is publicly accessible and free, with no profit being made from its content.
If you believe that any content on this blog infringes on your copyright, please contact me at the following email address, and I will promptly remove the content in question after verification.
Thank you for your understanding and support.
Contact: kechang.dev@gmail.com

A short history of AI 一则简短的 AI 历史

IMAGE: MIKE HADDAD
IMAGE: MIKE HADDAD
Jul 16th 2024
 
Over the summer of 1956 a small but illustrious group gathered at Dartmouth College in New Hampshire; it included Claude Shannon, the begetter of information theory, and Herb Simon, the only person ever to win both the Nobel Memorial Prize in Economic Sciences awarded by the Royal Swedish Academy of Sciences and the Turing Award awarded by the Association for Computing Machinery. They had been called together by a young researcher, John McCarthy, who wanted to discuss “how to make machines use language, form abstractions and concepts” and “solve kinds of problems now reserved for humans”. It was the first academic gathering devoted to what McCarthy dubbed “artificial intelligence”. And it set a template for the field’s next 60-odd years in coming up with no advances on a par with its ambitions.
1956 年夏天,一个杰出的小团体聚集在 New Hampshire 的 Dartmouth College;包含信息论的创始人 Claude Shannon,以及至今唯一一位同时获得瑞典皇家科学院授予的经济学的 Nobel Memorial Prize 和 由 ACM 授予 Turing Award 的 Herb Simon。他们被一个年轻的学者 John McCarthy 号召在一起,这个年轻的学者想要讨论 “如何让机器使用语言、形成抽象和概念” 和 “解决目前人类所保留的问题”。这是第一个专门讨论 McCarthy 所称的 “人工智能” 的学术聚会,为该领域接下来的大约六十多年奠定了模板,即使这些年未能取得与其雄心相匹配的突破性进展。 over the summer of 表示在某个特定年份的夏天期间发生的事情。它通常用来描述一段时间内的活动、变化或事件。 illustrious adj. 著名的;杰出的 begetter n. 生产者 swedish adj. 瑞典的;瑞典人的;瑞典语的 | n. 瑞典话; 瑞典人 machinery n. 机械装置;机械(总称) dub v. 起绰号;把…称为;配音 devoted to phrase. 致力于 -odd 表示大约的数量,不精确但接近
 
In this series on artificial intelligence
  1. A short history of AI*
  1. Running out of data
  1. Controlling the chip supply
 
The Dartmouth meeting did not mark the beginning of scientific inquiry into machines which could think like people. Alan Turing, for whom the Turing prize is named, wondered about it; so did John von Neumann, an inspiration to McCarthy. By 1956 there were already a number of approaches to the issue; historians think one of the reasons McCarthy coined the term artificial intelligence, later AI, for his project was that it was broad enough to encompass them all, keeping open the question of which might be best. Some researchers favoured systems based on combining facts about the world with axioms like those of geometry and symbolic logic so as to infer appropriate responses; others preferred building systems in which the probability of one thing depended on the constantly updated probabilities of many others.
Dartmouth 的会议并没有标志着对能够像人一样思考的机器的科学研究的开端。 Turing prize 名称的来源 Alan Turing 也疑惑过。McCarthy 的灵感来源 John von Neumann 同样如此。在 1956 年,以及有了一些方案可以解决这个问题。历史学家认为 McCarthy 为他的项目创造人工智能(之后称 AI)这个术语的其中一个原因是,它足够宽泛,可以涵盖所有方法,而哪种方法可能最好则没有答案。一些研究人员倾向于将公理与几何和符号逻辑等公理相结合的系统,以便推断出适当的结论。其他人则倾向于建立这样的系统:一件事的概率取决于许多其他事情不断更新的概率。 so did (So + 助动词(do/does/did)+ 主语)表示过去的动作或状态同样发生在另一个主体上 historian n. 历史学家 encompass v. 包围; 包含 keeping open the question 表示继续保持某个问题或争议的开放状态,暂时不作出最终决定或结论 axiom n. 公理;自明之理 like those of 说明一个事物与另一个事物在某些方面相似,尤其是指代前面提到的多个事物的特征或属性 so as to + v. 是用来表达行动或行为的目的,相当于“为了”或“以便于”的意思
 
Source: Stanford University AI Index report 2024
Source: Stanford University AI Index report 2024
The following decades saw much intellectual ferment and argument on the topic, but by the 1980s there was wide agreement on the way forward: “expert systems” which used symbolic logic to capture and apply the best of human know-how. The Japanese government, in particular, threw its weight behind the idea of such systems and the hardware they might need. But for the most part such systems proved too inflexible to cope with the messiness of the real world. By the late 1980s AI had fallen into disrepute, a byword for overpromising and underdelivering. Those researchers still in the field started to shun the term.
在接下来的几十年里,这个话题经历了许多智慧上的动荡和争论,但到了 20 世纪 80 年代,人们对于前进的方向达成了广泛的共识:“专家系统”使用符号逻辑来捕捉和应用人类最好的知识。日本政府尤其大力支持此类系统的想法以及它们可能需要的硬件。但在大多数情况下,这些系统应对现实世界的混乱都过于死板。到 20 世纪 80 年代末,人工智能已声名狼藉,成为言过其实、交付不足的代名词。那些仍在该领域的研究人员开始回避这个词。 ferment n. 发酵;酵素;动乱 | v. 使发酵;动乱 threw its weight behind 表示某个组织、团体、或个人积极支持或大力支持某个想法、计划、提议或某个特定的行动 cope v. 应付;处理 messiness n. 杂乱状态 disrepute n. 丧失名誉 byword n. 代名词; 笑柄;谚语 overpromising 言过其实 underdelivering 交付不足 shun v. 避开
 
It was from one of those pockets of perseverance that today’s boom was born. As the rudiments of the way in which brain cells—a type of neuron—work were pieced together in the 1940s, computer scientists began to wonder if machines could be wired up the same way. In a biological brain there are connections between neurons which allow activity in one to trigger or suppress activity in another; what one neuron does depends on what the other neurons connected to it are doing. A first attempt to model this in the lab (by Marvin Minsky, a Dartmouth attendee) used hardware to model networks of neurons. Since then, layers of interconnected neurons have been simulated in software.
今天的繁荣起源于那些顽强坚持的小群体。20 世纪 40 年代,随着脑细胞(一种神经元)工作方式的雏型被拼凑起来,计算机科学家开始怀疑是否可以使机器以同样的方式联结起来。在生物大脑中,神经元之间存在连接,这使得一个神经元的活动能够触发或抑制另一个神经元的活动。一个神经元的活动取决于与其相连的其他神经元的活动。首次在实验室中的尝试(由 Dartmouth 的出席者 Marvin Minsky 完成)使用硬件来模拟神经网络。此后,互连神经元层开始在软件中模拟。 perseverance n. 毅力;坚持不懈;不屈不挠 rudiment n. 基本原理;雏型 neuron n. 神经元 trigger n. 触发;触发器;板机;诱因 |v. 触发;引起;发动 attendee n. (会议)出席者 interconnecte v. 相互连接的;互相联系的 | v. (使)相互连接;(使)互相联系 simulate v. 假装;冒充;模仿;模拟
 
These artificial neural networks are not programmed using explicit rules; instead, they “learn” by being exposed to lots of examples. During this training the strength of the connections between the neurons (known as “weights”) are repeatedly adjusted so that, eventually, a given input produces an appropriate output. Minsky himself abandoned the idea, but others took it forward. By the early 1990s neural networks had been trained to do things like help sort the post by recognising handwritten numbers. Researchers thought adding more layers of neurons might allow more sophisticated achievements. But it also made the systems run much more slowly.
这些人工神经网络不使用明确的规则进行编程的,而是通过接触大量示例来“学习”。以便于在训练过程中,神经元之间的连接强度(称为“权重”)反复调整,最终对给定的输入产生适当的输出。Minsky 本人放弃了这个想法,但其他人将其发扬光大。到 20 世纪 90 年代初,神经网络已经经过训练,能够通过识别手写数字来帮助对文章进行分类。研究人员认为,增加更多层神经元可能会实现更复杂的效果,但也会使系统运行速度大大降低。 so that 以便于 take it forward 某个已经开始的事情进一步推进或发展。它强调的是对某项工作或项目的延续、改进、或提升,是在已有基础上的进一步动作。 handwrite v. 手写 sophisticated adj. 老于世故的;老练的;精密复杂的
 
A new sort of computer hardware provided a way around the problem. Its potential was dramatically demonstrated in 2009, when researchers at Stanford University increased the speed at which a neural net could run 70-fold, using a gaming PC in their dorm room. This was possible because, as well as the “central processing unit” (cpu) found in all pcs, this one also had a “graphics processing unit” (gpu) to create game worlds on screen. And the gpu was designed in a way suited to running the neural-network code.
一种新型计算机硬件围绕这个问题提供了解决方案。2009 年,Stanford 大学的研究人员利用宿舍里的一台游戏电脑,将神经网络的运行速度提高了 70 倍,这一发现极大地证明了这种硬件的潜力。这是因为,除了所有电脑都有的“中央处理器”(cpu)外,这款电脑还配备了“图形处理单元”(gpu),可以在屏幕上创建游戏世界。并且,gpu 的设计方式适合运行神经网络代码。 -fold 用于描述数量或程度的倍数增加或减少。这个后缀的用法可以表达增长、扩展或减少的倍数 dorm n. (集体)宿舍
 
Coupling that hardware speed-up with more efficient training algorithms meant that networks with millions of connections could be trained in a reasonable time; neural networks could handle bigger inputs and, crucially, be given more layers. These “deeper” networks turned out to be far more capable.
将硬件加速与更高效的训练算法相结合,意味着拥有数百万个连接的网络可以在合理的时间内完成训练;神经网络可以处理更大的输入,而且最重要的是,可以被给予更多的层数。这些“更深层”的网络被证明更加强大。 coupling n. 联结;结合 | v-ing. 联结;结合 turned out to be 结果是;最后证明是
 
The power of this new approach, which had come to be known as “deep learning”, became apparent in the ImageNet Challenge of 2012. Image-recognition systems competing in the challenge were provided with a database of more than a million labelled image files. For any given word, such as “dog” or “cat”, the database contained several hundred photos. Image-recognition systems would be trained, using these examples, to “map” input, in the form of images, onto output in the form of one-word descriptions. The systems were then challenged to produce such descriptions when fed previously unseen test images. In 2012 a team led by Geoff Hinton, then at the University of Toronto, used deep learning to achieve an accuracy of 85%. It was instantly recognised as a breakthrough.
这种新方法被称为“深度学习”,其威力在 2012 年的 ImageNet 挑战赛中初露锋芒。参赛的图像识别系统被提供了一个包含超过一百万个带标签图像文件的数据库。对于任何给定的单词,例如 “dog” 或 “cat”,数据库中都包含数百张照片。图像识别系统将使用这些示例进行训练,将以图像形式的输入“映射”到以单词描述形式的输出上。然后,系统被要求在输入之前从未见过的测试图像时生成这样的描述。2012 年,Toronto 大学的 Geoff Hinton 领导的团队使用深度学习实现了 85% 的准确率。它立即被认为是一项突破。 unseen adj. 看不见的;未被发现的;缺席的
 
By 2015 almost everyone in the image-recognition field was using deep learning, and the winning accuracy at the ImageNet Challenge had reached 96%—better than the average human score. Deep learning was also being applied to a host of other “problems…reserved for humans” which could be reduced to the mapping of one type of thing onto another: speech recognition (mapping sound to text), face-recognition (mapping faces to names) and translation.
到了 2015 年,在图像识别领域几乎都使用了深度学习,并且 ImageNet 挑战赛中的胜者有着高达 96% 的准确度——超过了人类得分的平均水平。深度学习也被应用于许多其他“人类保留问题”,这些问题可以归结为将一种事物映射到另一种事物上:语音识别(将声音映射到文本)、人脸识别(将人脸映射到名字)和翻译。 a host of 它通常用于描述数量众多的某类事物或人,带有强调数量庞大的意味 be reduced to 减少到; 归结为
 
In all these applications the huge amounts of data that could be accessed through the internet were vital to success; what was more, the number of people using the internet spoke to the possibility of large markets. And the bigger (ie, deeper) the networks were made, and the more training data they were given, the more their performance improved.
在所有这些应用中,通过互联网获取的大量数据对于成功至关重要;此外,使用互联网的人数表明了巨大市场的可能性。并且,网络越大(即越深),获得的训练数据越多,其性能就越好。 what was more 更重要的是; 更有甚者; 而且 the more…the more 表达两者之间的相互关系或变化。这个结构表示随着一个事情(或状态)增加或增强,另一个事情(或状态)也随之增加或增强。通常用于描述两个变量之间的正相关关系
 
Deep learning was soon being deployed in all kinds of new products and services. Voice-driven devices such as Amazon’s Alexa appeared. Online transcription services became useful. Web browsers offered automatic translations. Saying such things were enabled by AI started to sound cool, rather than embarrassing, though it was also a bit redundant; nearly every technology referred to as AI then and now actually relies on deep learning under the bonnet.
深度学习很快应用于各种新产品和服务中。Amazon 的 Alexa 等语音驱动设备问世。在线转录服务变得实用。网络浏览器提供自动翻译。说这些项目由 AI 赋能听起来很酷,而不是令人尴尬,尽管这也有点多余;当时和现在,几乎所有被称为人工智能的技术实际上都依赖于深度学习。 transcription n. 抄写;誊写 redundant adj. 多余的;过剩的 bonnet n. 软帽;阀帽;引擎盖 under the bonnet 在内部; 在背后
 
Chatgpt and its rivals really do seem to “use language and form abstractions” ChatGPT及其竞争对手确实似乎能够 “使用语言并形成抽象概念”
In 2017 a qualitative change was added to the quantitative benefits being provided by more computing power and more data: a new way of arranging connections between neurons called the transformer. Transformers enable neural networks to keep track of patterns in their input, even if the elements of the pattern are far apart, in a way that allows them to bestow “attention” on particular features in the data.
2017 年,更强大的计算能力和更多的数据所带来的数量效益发生了质的变化:一种安排神经元之间连接的新方法,称为“Transformer”。Transformer 使神经网络能够跟踪输入中的模式,即使模式中的元素相距很远,也能以某种方式让它们对数据中的特定特征给予“注意力”。 rival n. 竞争者;敌手 | v. 竞争 qualitative adj. 性质的;定性的 quantitative adj. 数量的;定量的 qualitative change 质变 quantitative benefits 量变效应 keep track of 追踪;纪录 bestow v. 给予;赠予;授予 bestow … on phrase. 给予;赠予;授予
 
Transformers gave networks a better grasp of context, which suited them to a technique called “self-supervised learning”. In essence, some words are randomly blanked out during training, and the model teaches itself to fill in the most likely candidate. Because the training data do not have to be labelled in advance, such models can be trained using billions of words of raw text taken from the internet.
Transformer 让神经网络更好抓住上下文,这让它们更适合一项称为“自监管学习”的技术。本质上,在训练过程中,一些单词会被随机删除,然后模型会自学填写最有可能的候选单词。由于训练数据不需要提前标记,因此这些模型可以使用从互联网上获取的数十亿字的原始文本来训练。 in advance phrase. 提前 raw adj. 生的;未经处理的
 

Mind your language model 关注你的语言模型

Transformer-based large language models (LLMs) began attracting wider attention in 2019, when a model called GPT-2 was released by OpenAI, a startup (GPT stands for generative pre-trained transformer). Such LLMs turned out to be capable of “emergent” behaviour for which they had not been explicitly trained. Soaking up huge amounts of language did not just make them surprisingly adept at linguistic tasks like summarisation or translation, but also at things—like simple arithmetic and the writing of software—which were implicit in the training data. Less happily it also meant they reproduced biases in the data fed to them, which meant many of the prevailing prejudices of human society emerged in their output.
基于 Transformer 的大语言模型在 2019 年开始受到广泛关注,当时一个由初创公司 OpenAI 发布了名为 GPT-2(GPT 代表 生成式预训练 Transformer) 的模型。这些 LLM 能够表现出 “emergent” 行为,尽管他们曾经没有接受过专门的训练。吸收大量的语料不仅使它们出人意料地擅长语言学任务(比如总结或翻译),而且还擅长训练数据中隐含的事情(比如简单的算术和软件编写)。但不幸的是,这也意味着它们在输入的数据中重现了偏见,这意味着人类社会许多普遍的偏见都出现在它们的输出中。 startup n. 启动;新兴公司;初创公司 stand for phrase. 代表;象征;支持;容忍 turn out to be phrase. 结果是;原来是;证实是 explicitly adv. 明白地;明确地;直言地 soak up phrase. 吸收;吸取 surprisingly adv. 令人吃惊地 adept adj. 内行的;熟练的;擅长的 | n. 内行 summarisation n. 总结 arithmetic n. 算术 implicit adj. 含蓄的;暗示的;无疑的;绝对的 prevailing adj. 普遍的;盛行的;有优势的 | v-ing. 获胜;盛行;劝服 prejudice n. 偏见;成见 | v. 使有偏见;损害
 
In November 2022 a larger OpenAI model, GPT-3.5, was presented to the public in the form of a chatbot. Anyone with a web browser could enter a prompt and get a response. No consumer product has ever taken off quicker. Within weeks ChatGPT was generating everything from college essays to computer code. AI had made another great leap forward.
2022 年 11 月,更大的 OpenAI 模型 GPT-3.5 以聊天机器人的形式向公众公开。任何拥有网络浏览器的人都可以输入 prompt 并获得响应。从来没有任何消费产品能够像它一样迅速流行。发布后的短短几周内,ChatGPT 就展现出了它强大的生成能力,涵盖了从学术写作到编程等多个领域。人工智能又一次实现了巨大的飞跃。 take off phrase. 突然成功;拿掉;离开;起飞;脱下 within weeks 短短几周内 another great leap forward 又一次巨大的飞跃
 
Where the first cohort of AI-powered products was based on recognition, this second one is based on generation. Deep-learning models such as Stable Diffusion and DALL-E, which also made their debuts around that time, used a technique called diffusion to turn text prompts into images. Other models can produce surprisingly realistic video, speech or music.
第一批 AI 赋能产品基于识别,而第二批则基于生成。Stable Diffusion 和 DALL-E 等深度学习模型也是在那时首次亮相,它们使用一种称为 diffusion 的技术将文本 prompt 转换为图像。其他模型可以制作出令人惊讶的逼真的视频、语音或音乐。 cohort n. 一群人;一帮人 debut n. 初次露面;初次登场 | v. 初次登场;首次推出 around that time 在那个时候;大约在同一时间 diffusion n. 扩散;传播;漫散
 
The leap is not just technological. Making things makes a difference. ChatGPT and rivals such as Gemini (from Google) and Claude (from Anthropic, founded by researchers previously at OpenAI) produce outputs from calculations just as other deep-learning systems do. But the fact that they respond to requests with novelties makes them feel very unlike software which recognises faces, takes dictation or translates menus. They really do seem to “use language” and “form abstractions”, just as McCarthy had hoped.
这种飞跃不仅仅是技术上的。创造能够带来改变。ChatGPT 以及它的竞争对手,例如来自 Google 的 Gemini 和 Anthropic(由前 OpenAI 研究人员创立的公司)的Claude 通过计算生成输出,这也像其他深度学习系统一样。但它们通过新颖的方式回应请求这一事实,使它们看起来非常不同于那些识别人脸、进行语音转译或翻译菜单的软件。它们似乎确实在 “使用语言” 和 “形成抽象概念”,正如 McCarthy 所希望的那样。 Making things makes a difference. 创造能够带来改变。
 
This series of briefs will look at how these models work, how much further their powers can grow, what new uses they will be put to, as well as what they will not, or should not, be used for. ■
本系列简报将探讨这些模型是怎样工作的、它们的能力还能增强到何种程度、它们将用于何处以及它们不会或不应该用于什么用途。■
 
AI firms will soon exhaust most of the internet’s data AI 公司将很快用尽互联网中的数据“Black Myth: Wukong” is China’s first blockbuster video game 中国首部重磅游戏:“黑神话:悟空”
Loading...