Richard Feynman must have been one of those rare visionaries who predict the future based not on mere creativity of the mind, but on creativity firmly entrenched in subject matter expertise. To him, miniaturization of processes and tools was a logical next step to resolve the shortcomings of existing technology – such as the slowness and bulkiness of computers in the 1950s. While most avant-garde scientific thinkers would have stopped at imagining miniaturization at micro-levels, Feynman went a step further – why not explore the molecular scale of matter? Decades before the term “nanotechnology” was coined and popularized by Eric Drexler, Feynman’s seminal 1959 lecture “Plenty of Room at the Bottom” presented the concepts of nano-scale miniaturization as a logically inevitable development in computing, chemistry, engineering, and even medicine.
Feynman’s timescale predictions were at once overly optimistic and overly pessimistic – on the one hand, he expected nanoscale miniaturization of, say, data storage to be a reality by the year 2000, and already feasible during the early 1970s. On the other hand, some of the applications which he believed would call for a sub-microscopic scale – such as computers capable of face recognition – exist already, and are widely implemented by law enforcement and social networks alike.
As for Feynman’s idea of writing the entire human knowledge onto media as small as pinheads, to the tune of taking up the print size of a pamphlet (readable through an electron microscope improved by a factor of 100), it is not the size of the media that currently poses the problem: the issue humanity is facing is, firstly, the exploding amount of its aggregate knowledge, and, secondly, the challenge of its digitization. In 1959, according to Feynman, the Library of Congress housed 9 million volumes. Today, its holdings include 151 million items in 470 languages, with millions added each year. The vast majority of all this information has not yet been digitized – and here we come to the second problem. It takes thousands of human users to transfer old books into digital form, and the pace of the volunteer-run Project Gutenberg shows how time-consuming the current practice of transcribing books really is, even when aided by continually advancing OCR software. Since 1971, Project Gutenberg digitized a mere 42,000 books in the public domain. Crowdsourcing promises a possible quantum leap in the acceleration of knowledge digitization, with reCAPTCHA (an ingenuous reverse application of CAPTCHA, the technology used to authenticate human users online by requiring them to identify distorted signs or words as a condition of access to a certain software function or service) harnessing involuntarily the resources of 750 million computer users world-wide to digitize annually 2.5 million books that cannot be machine-read. This solution, by the way, is just another development imagined by Feynman way ahead of his time: parallel computing.
As is quantum computing. Even though Richard Feynman announced the novel idea of quantum computers only in the 1980s, the first (and very rudimentary) models of simple quantum machines are already in existence today.
Almost all of Feynman’s groundbreaking ideas are slowly becoming everyday reality. We may not have usable nanocomputers just yet, but nanotechnology and nanomaterials are slowly but steadily entering the world as we know it. And as for the writing of entire encyclopedias on headpins, “just for fun”? The smallest book in the Library of Congress today is “Old King Cole” measuring 1/25” x 1/25”, or 1 sq mm. That is the size of a period at the end of the sentence.
Post a Comment