Converging Cognitive Enhancements
Nick Bostrom
2006-12-12 00:00:00
URL

CONVERGING COGNITIVE ENHANCEMENTS

There are few resources more useful than cognitive ability. While other resources are necessary or desirable, cognition enables them to be used for achieving personal goals. While there is little evidence that high intelligence causes happiness there appears to be ample evidence that low intelligence increases the risk for accidents, negative life events, and low income (Gottfredson 1997, 2004) while higher intelligence promotes health (Whalley and Deary 2001) and wealth. We also need better cognition in order to balance an increasingly complex society where information becomes more available and our actions have more far-reaching consequences (Heylighen 2002a, 2002b). There may also be an intrinsic existential value in being able to perceive, understand, and interact well with the world.

Cognitive enhancement may be defined as the amplification or extension of core capacities of the mind through improvement or augmentation of internal or external information processing systems. Cognition in turn can be defined as the processes an organism uses to organize information. This includes both the acquisition of information (perception), selecting (attention), representing (understanding), and retaining (memory) information, and using it to guide behavior (reasoning and coordination of motor outputs). Interventions to improve cognitive function may be directed at any one of these core faculties.

As cognitive neuroscience has advanced, the list of prospective internal, biological enhancements has steadily expanded (Farah et al. 2004). Yet to date, it is progress in information technology and cultural organization that has produced the most dramatic advances in our ability to process information. External hardware and software supports now routinely give human beings effective cognitive abilities that in many respects far outstrip those of our native minds, and institutions like peer review or markets.

There exists a long tradition in human–computer interaction dealing with cognitive enhancement, beginning with William Ross Ashby defining intelligence as the “power of appropriate selection,” which could be technologically amplified similar to physical power (Ashby 1956). By offloading mental tasks to computers or embedding humans within a software context their cognitive functioning could be amplified (Licklider 1960). The aim was not artificial intelligence but rather amplifying human intelligence. The archetypal example of this approach is Douglas C. Engelbart’s famous Augmenting Human Intellect, which defined the goal as:

By ‘augmenting human intellect’ we mean increasing the capability of a man to approach a complex problem situation, to gain comprehension to suit his particular needs, and to derive solutions to problems. Increased capability in this respect is taken to mean a mixture of the following: more-rapid comprehension, better comprehension, the possibility of gaining a useful degree of comprehension in a situation that previously was too complex, speedier solutions, better solutions, and the possibility of finding solutions to problems that before seemed insoluble. And by ‘complex situations’ we include the professional problems of diplomats, executives, social scientists, life scientists, physical scientists, attorneys, designers—whether the problem situation exists for twenty minutes or twenty years. We do not speak of isolated clever tricks that help in particular situations. We refer to a way of life in an integrated domain where hunches, cut-and-try, intangibles, and the human ‘feel for a situation’ usefully co-exist with powerful concepts, streamlined terminology and notation, sophisticated methods, and high-powered electronic aids.

Man’s population and gross product are increasing at a considerable rate, but the complexity of his problems grows still faster, and the urgency with which solutions must be found becomes steadily greater in response to the increased rate of activity and the increasingly global nature of that activity. Augmenting man’s intellect, in the sense defined above, would warrant full pursuit by an enlightened society if there could be shown a reasonable approach and some plausible benefits (Engelbart 1962).

An important insight was that it is not enough to improve just computer hardware and software, but psychological and organizational aspects have to be taken into account.

The cybernetic approach has in itself been technology independent by focusing on what is enhanced rather than the means of doing it. This unfortunately also causes disconnection from the richer social–ethical debate surrounding the other approaches, because they mostly take place within bioethics and medical ethics.

Studying cognitive enhancement solely in terms of bioethics, computer supported intelligence amplification or nanomedicine, risks missing the key commonalities. Converging technologies give a framework to approach the commonalities between different forms of human enhancement, as well as a way to contrast their differences and potential for divergence.

Criticisms of enhancements are often stated in a technology-independent form yet when analyzed from a converging technologies perspective they often show strong assumptions about a particular kind of technology. Those that are truly technology independent, even if originating within in a narrow area such as the genetics discourse, on the other hand raise relevant challenges for broad areas.


PHARMACEUTICAL BIOTECHNOLOGY

Today there exist a broad range of drugs that can affect cognition. Stimulant drugs like nicotine and caffeine are traditionally and widely used to improve cognition. In the case of nicotine a complex interaction with attention and memory occurs (Warburton 1992; Newhouse et al. 2004; Rusted et al. 2005) while caffeine reduces tiredness (Lieberman 2001; Smith et al. 2003; Tieges et al. 2004).

Lashley observed in 1917 that strychnine facilitates learning in rats (Lashley 1917). Since then several families of memory-enhancing drugs affecting different aspects of long-term memory have been discovered. They range from stimulants (Soetens et al. 1993; Lee and Ma 1995; Soetens et al. 1995), nutrients (Foster et al. 1998; Korol and Gold 1998; Winder and Borrill 1998; Meikle et al. 2005), and hormones (Gulpinar and Yegen 2004) over cholinergic agonists (Iversen 1998; Power et al. 2003; Freo et al. 2005) and the piracetam family (Mondadori 1996) to ampakines (Ingvar et al. 1997; Lynch 1998) and consolidation enhancers (Lynch 2002). The earliest drugs were mainly nonspecific stimulants and nutrients. For example, during antiquity honey water, hydromel, was used for doping purposes. Glucose is the major energy source for the brain, which relies on a continuous supply to function. Increases in availability (either due to ingestion or stress hormones) improve memory (Wenk 1989; Foster et al. 1998). Stimulants enhance either by increasing the amount of neuron activity or by releasing neuromodulators, both factors which make the synaptic change underlying learning more likely.

The growing understanding of memory allowed the development of more specific drugs. Stimulating the cholinergic system, which appears to gate attention and memory encoding, was a second step. Current interest is focused on intervening into the process of permanent encoding in the synapses, which has been elucidated to a great extent and hence has become a promising target for drug development. The goal would be drugs that not just allow the brain to learn quickly but also facilitate selective retention of the information that has been learned. It is known that the above families of drugs can improve performance in particular memory tests. It is not yet known whether they also promote useful learning in real-life situations.

Pharmacological agents might be useful not only for increasing memory retention, but also for unlearning phobias and addictions (Pitman et al. 2002; Ressler et al. 2004; Hofmann et al. 2006). Potentially, the combination of different pharmacological agents administered at different times could allow users a more fine-grained control of their learning processes, and perhaps even the ability to deliberately select the contents of their memory.

Even common, traditional, and unregulated herbs and spices, such as sage, can improve memory and mood through chemical effects (Kennedy et al. 2006). While less powerful than those of dedicated cholinesterase inhibitors, such effects illustrate that attempts to control access to cognition-enhancing substances would be problematic. Even chewing gum appears to affect memory, possibly by heightening arousal or blood sugar (Wilkinson et al. 2002).

Working memory can be modulated by a variety of drugs. Drugs that stimulate the dopamine system have demonstrated effects, as do cholinergic drugs (possibly through improved encoding) (Barch 2004). Modafinil has been shown to enhance working memory in healthy test subjects, especially at harder task difficulties and for lower performing subjects (Muller et al. 2004). (Similar findings, of greater improvements among low performers were also seen among the dopaminergic drugs, and this might be a general pattern for many cognitive enhancers.) On a larger battery of tasks, modafinil was found to increase forward and backward digit span, visual pattern recognition memory, spatial planning, and reaction time/latency on different working memory tasks (Turner et al. 2003). The reason might be that modafinil enhances adaptive response inhibition, making the subjects evaluate a problem more thoroughly before responding to it, thereby improving performance accuracy. The working memory effects might hence be part of a more general enhancement of executive function. A few other drugs may also improve executive function (Elliott et al. 1997; Kimberg et al. 1997; Mehta et al. 2000). Given that these functions are closely linked to what is commonly seen as intelligence, they may be the first step toward true intelligence-enhancing drugs.

Modafinil was originally developed as a treatment for narcolepsy, and can be used to reduce the performance decrements due to sleep loss with apparently small side effects and risk of dependency (Teitelman 2001; Myrick et al. 2004). The drug improved attention and working memory in sleep-deprived physicians (Gill et al. 2006) and aviators (Caldwell et al. 2000). Naps are more effective in maintaining performance than modafinil and amphetamine during long (48 h) periods of sleep deprivation than during short (24 h), but naps followed by a modafinil dose may be more efficient than either individually (Batejat and Lagarde 1999). These results, together with hormones like melatonin that can control sleep rhythms (Cardinali et al. 2002), suggest that drugs can help shape sleep and alertness patterns to improve task performance under demanding circumstances.

Creativity can also be affected pharmacologically. A study using alcohol demonstrated that a mild dose of alcohol could improve the results of a creative scientific process (Norlander and Gustafson 1996). The improvement only occurred when the subjects got the alcohol during the “incubation phase” of the creative process, the period when they were not actively working on the problem but presumably their unconscious might have been active. Giving alcohol in a picture-drawing task during the later verification phase did not promote creativity (Norlander and Gustafson 1997).

Creative thinking does not just include divergent and disinhibited thinking, but also requires convergent thinking to focus on the realization of the insight (Cropley 2006). Excessive divergence or lack of inhibition may be similar to the situation in attention deficit hyperactivity disorder (ADHD). Adult ADHD individuals show a profile of divergent thinking and do badly on convergent thinking and inhibition tasks (White and Shah 2006). Hence medications affecting ADHD might promote convergent thinking. Methylphenidate, the most common treatment and a potential executive function enhancer, did not appear to impair flexible thinking in ADHD individuals (Solanto and Wender 1989; Douglas et al. 1995). Giving L-dopa, a dopamine precursor, to healthy volunteers did not affect direct semantic priming (faster recognition of words directly semantically related to a previous word, such as “black-white”) but did inhibit indirect priming (faster recognition of more semantically distant words, such as “summer-snow”) (Kischka et al. 1996). This was interpreted by the authors of the study as dopamine inhibiting the spread of activation within the semantic network, that is, a focusing on the task.

There also exist drugs that influence how the cerebral cortex reorganizes in response to damage or training. Noradrenergic agonists, such as amphetamine, have been shown to promote faster recovery of function after a brain lesion when combined with training (Gladstone and Black 2000) and to improve learning of an artificial language (Breitenstein et al. 2004). A likely explanation is that higher excitability increases cortical plasticity, in turn leading to synaptic sprouting and remodeling (Stroemer et al. 1998; Goldstein 1999). An alternative to pharmacological increase of neuromodulation is to electrically stimulate the neuromodulatory centers that normally control plasticity through attention or reward. In monkey experiments this produced faster cortical reorganization (Kilgard and Merzenich 1998; Bao et al. 2001).

In general, pharmacological enhancement is possible here and now, although the improvements in ability tend to be a modest 10–20% improvement of test scores. As for all pharmacology, there are great interindividual variations. Using enhancer drugs optimally might include tests of neuromodulator levels to see where the brain setpoints are, pharmacogenomic tests to find how they are metabolized and neuropsychological tests to check what levels produce maximum performance. Such fine-tuning is expensive and cumbersome unless it can be automated.


OTHER BIOTECHNOLOGIES

Transcranial magnetic stimulation (TMS) stimulates neurons in the cerebral cortex bya changing magneticfield inducedfrom a coil heldtothehead.Itcan increase or decrease the excitability of the cortex, thereby changing its level of plasticity (Hummel and Cohen 2005). TMS of the motor cortex that increased its excitability improved performance in a procedural learning task (Pascual- Leone et al. 1999). TMS in suitable areas has also been found beneficial in a motor task (Butefisch et al. 2004), motor learning (Nitsche et al. 2003), visuomotor coordination tasks (Antal et al. 2004a, 2004b), working memory (Fregni et al. 2005), finger sequence tapping (Kobayashi et al. 2004), classification (Kincses et al. 2004), and even declarative memory consolidation during sleep (Marshall et al. 2004). Snyder et al. demonstrated how TMS inhibiting anterior brain areas could change the drawing style of normal subjects into a more concrete style and improve spell-checking abilities, presumably by reducing top-down semantic control (Snyder et al. 2003, 2004). While TMS appears to be highly versatile and noninvasive, there are risks of triggering epileptic seizures and the effects of long-term use are not known. Individual brain differences may necessitate much adjustment before it can be applied to a specific use.

Genetic memory enhancement has been demonstrated in rats and mice. In normal animals, during maturation expression of the NR2B subunit of the Nmethyl- D-aspartate (NMDA) receptor is gradually replaced with expression of the NR2A subunit, something that may be linked to the lower brain plasticity in adult animals. Tsien’s group (Tang et al. 1999) modified mice to overexpress the NR2B. The NR2B “Doogie” mice demonstrated improved memory performance, both in terms of acquisition and retention. This included unlearning of fear conditioning, which is believed to be due to the learning of a secondary memory (Falls et al. 1992). The modification also made the mice more sensitive to certain forms of pain, suggesting a nontrivial trade-off between two potential enhancement goals (Wei et al. 2001).

Increased amounts of brain growth factors (Routtenberg et al. 2000) and the signal transduction protein adenylyl cyclase (Wang et al. 2004) have also produced memory improvements. These modifications have different enhancing effects: unlearning took longer for these modified mice than for unmodified mice, while the mice in the Tsien study had faster than normal unlearning. Different memory tasks were also differently affected: the cyclase mice had enhanced recognition memory but not improved context or cue learning. A fourth study showed that mice with a deleted cbl-b gene had normal learning but enhanced long-term retention, presumably indicating that the gene is a negative regulator of memory (Tan et al. 2006). These enhancements may be due to changes in neural plasticity during the learning task itself, or that the developing modified brain develops in a way that promotes subsequent learning or retention.

The cellular machinery of memory appears to be highly conserved in evolution, making interventions demonstrated to work in animal models likely to have close counterparts in humans (Edelhoff et al. 1995; Bailey et al. 1996).

Genetic studies have also found genes in humans whose variations account for up to 5% of memory performance (de Quervain and Papassotiropoulos 2006). These include the genes for the NMDA receptor and adenylyl cyclase that were mentioned above, as well as other parts of the synaptic signal cascade. These are clear targets for enhancement.

Given these early results, it seems likely that there exist many potential genetic interventions that directly or indirectly improve aspects of memory. If it turns out that the beneficial effects of the treatments are not due to changes in development, then presumably some of the effects can be achieved by supplying the brain with the substances produced by the memory genes without resorting to genetic modification. But genetic modification would make the individual independent of an external drug supply and would guarantee that the substances end up in the right place.

On the other hand, studies of the genetics of intelligence suggests that there is a large number of genetic variations affecting individual intelligence, but each accounting for only a very small fraction (<1%) of the variance between individuals (Craig and Plomin 2006). This would indicate that genetic enhancement of intelligence through direct insertion of a few beneficial alleles is unlikely to have a big enhancing effect. It is possible, however, that some alleles that are rare in the human population could have larger effects on intelligence, either negative or positive. A possible example is the prediction that heterozygoticity for Tay-Sachs’ disease should increase IQ by about 5 points (Cochran et al. 2006).

While human germline engineering is controversial, several years away and likely to be expensive, the genetic discoveries discussed here may be used for enhancement in other ways. Gene expression may be affected pharmacologically or even through food intake.

A notable form of chemical enhancement is pre-and perinatal enhancement through maternal nutrition. Administering choline supplementation to pregnant rats improved the performance of their pups, apparently as a result of changes in neural development in turn due to changes in gene expession (Meck et al. 1988; Meck and Williams 2003; Mellott et al. 2004). Given the ready availability of choline supplements, such prenatal enhancement, may already (inadvertently) be taking place in human populations. Supplementation of a mother’s diet during late pregnancy and 3 months postpartum with long-chained fatty acids has also been demonstrated to improve cognitive performance in human children (Helland et al. 2003). Deliberate changes of maternal diet may hence be seen as part of the cognitive enhancement spectrum. At present, recommendations to mothers are mostly aimed at promoting a diet that avoids specific harms and deficits, but the growing emphasis on boosting “good fats” and the use of enriched infant formulas point toward enhancement.


COGNITIVE TECHNOLOGY

Education has many benefits beyond higher job status and salary. Longer education reduces the risks of substance abuse, crime, and many illnesses while improving quality of life, social connectedness, and political participation (Johnston 2004). There is also positive feedback between performance on cognitive tests, such as IQ tests and scholastic achievement, producing a 2.7 IQ point advantage per year of schooling (Winship and Korenman 1997). While education may be more of a social enhancement technology than a cognitive enhancement technology, it clearly has some potential for the latter.

Much of what we learn in school is “mental software” for managing various cognitive domains: mathematics, categories of concepts, language, and problem solving in particular subjects. This kind of mental software reduces our mental load through clever encoding, organization, or processing. Instead of memorizing arbitrarily large multiplication tables we compress the pattern of arithmetic relationships into simpler rules of multiplication, which in turn (among very ambitious students) can be organized into efficient mental calculation methods like the Trachtenberg system (Trachtenberg 2000). Such specific methods have a smaller range of applicability but can dramatically improve performance within a particular domain. They represent a form of crystallized intelligence, distinct from the fluid intelligence of general cognitive abilities and problem solving capacity (Cattell 1987). The relative ease and utility of improving crystallized intelligence and specific abilities have made them popular targets of internal and external software development. Cognitive enhancement attempts the more difficult challenge of improving fluid intelligence.

The challenge of improving education is perennial, and much hope is currently placed on using the results of neuroscience to improve education. However, so far pure neuroscience has provided few directly applicable tools (Goswami 2006). While this may change, the deep interdisciplinary divide that has to be bridged may prove a far greater challenge than most forms of technological convergence.

Pharmacological cognitive enhancements (nootropics) have physiological effects on the brain. So too do education and other conventional interventions. In fact, conventional interventions often produce more permanent neurological changes than do drugs. Learning to read alters the way language is processed in the brain (Petersson et al. 2000). Enriched rearing environments have been found to increase dendritic arborization and to produce synaptic changes, neurogenesis, and improved cognition in animals (Walsh et al. 1969; Greenough and Volkmar 1973; Diamond et al. 1975; Nilsson et al. 1999). While analogous controlled experiments cannot easily be done for human children, it is very likely that similar effects would be observed. Stimulation-seeking children, who might be seeking out and creating enriched environments for themselves, score higher on IQ tests and do better at school than less stimulation-seeking children (Raine et al. 2002). This also suggests that interventions that make exploring and learning more appealing to children, whether environmental or perhaps pharmaceutical, would have significant cognition-enhancing effects.

Enriched environments also make brains more resilient to stress and neurotoxins (Schneider et al. 2001). Reducing neurotoxins and preventing bad prenatal environments are simple and widely accepted methods of increasing cognitive function. These latter kinds of intervention might be classified as preventative or therapeutic rather than enhancing, although the distinction is blurry. For instance, an optimized intrauterine environment will not only help avoid specific pathology and deficits but is also likely to promote the growth of the developing nervous system in ways that ultimately enhance its core capacities.

In brains that have already been damaged, for example, by lead exposure, nootropics may alleviate some of the cognitive deficits (Zhou and Suszkiw 2004). It is not always clear whether they do this by curing the damage or by amplifying (enhancing) capacities that can compensate for the loss, or even whether the distinction is always meaningful. Comparing chronic exposure to cognition-enhancing drugs with an enriched rearing environment, one study found that both conditions improved memory performance and produced similar changes in the neural matter in rats. The improvements in the drug-treated group persisted even after cessation of treatment. The combination of drugs and enriched environment did not improve the rats’ abilities beyond the improvement provided by one of the interventions alone. This suggests that both interventions produced a more robust and plastic neural structure able to learn more efficiently.

Improving general health has cognition-enhancing effects. Many health problems act as distractors or directly impair cognition (Schillerstrom et al. 2005). Improving sleep, immune function, and general conditioning promotes cognitive functioning. Bouts of exercise have been shown to temporally improve various cognitive capacities, the size of the effect depending on the type and intensity of the exercise (Tomporowski 2003). Long-term exercise also improves cognition, possibly by a combination of increased blood supply to the brain and the release of nerve growth factors (Vaynman and Gomez-Pinilla 2005). Understanding this system may lead to new classes of nootropics, perhaps as a side effect of research into regenerative medicine.

Overall, improvements in the environment may be effective and widely acceptable cognition enhancers, and conversely enhancement may help deprived individuals.

MENTAL TRAINING

Mental training and visualization techniques are widely practiced in elite sport (Feltz and Landers 1983) and rehabilitation (Jackson et al. 2004), with apparently good effects. Users vividly imagine themselves performing a task (running a race, going to a store), imagining every movement and how they feel again and again. A likely explanation for the efficacy of such exercises is that they activate the neural networks involved in executing a skill at the same time as the performance criteria for the task is held in close attention, optimizing neural plasticity and appropriate neural reorganization/ learning.

Even general mental activity, “working the brain muscle” can improve performance (Nyberg et al. 2003) and long-term health (Barnes et al. 2004), as can relaxation techniques to regulate the activation of the brain (Nava et al. 2004). It has been suggested that the Flynn-effect (Flynn 1987), a secular increase in raw intelligence test scores by 2.5 IQ points per decade in most western countries, can be attributed to increased demands of certain forms of abstract and visuospatial cognition in modern society and schooling, although improved nutrition and health status may also play a part (Neisser 1997; Blair et al. 2005). It appears that most of the Flynn effect does not reflect an increase in general fluid intelligence but rather a change in which specific forms of intelligence are developed.

The classic form of cognitive enhancement software is learned strategies to memorize information. Such methods have been used since antiquity with much success (Yates 1966; Patten 1990). One such classic strategy is “the method of loci.” The user imagines a building, either real or imaginary, and in her imagination she walks from room to room, and places imaginary objects that evoke natural associations to the subject matter that she is memorizing. During retrieval, the user retraces her steps and the sequence of memorized information is recalled when she “sees” the objects she has placed along the route. This technique harnesses the brain’s spatial navigation system to help remember objects or propositional contents. Other memory techniques makes use of rhyming or the fact that we more easily recall dramatic, colorful, or emotional scenes, which can serve as placeholders for items that are more difficult to retain, such as numbers or letters. The early memory arts were often used as a substitute for written text or to memorize speeches. Today, memory techniques tend to be used in service of everyday needs, such as remembering door codes, passwords, shopping lists, and by students who need to memorize names, dates, and terms when preparing for exams (Lorrayne 1996; Minninger 1997).

One study which compared exceptional memorizers (participants in the World Memory Championships) with normal subjects found no systematic differences in brain anatomy (Maguire et al. 2003). However, activity during encoding was different, likely reflecting the use of a deliberate encoding strategy. Especially areas of the brain involved in spatial representation and navigation were found to be consistently activated in the memorizers, regardless of whether the subjects were learning numbers, faces, or snowflakes. When asked about their memory strategies, nearly all memorizers reported using the method of loci.

In general it appears possible to attain very high memory performance on specific types of material using memory techniques. They work best on otherwise meaningless or unrelated information, such as sequences of numbers, but do not appear to help skilled everyday activities (Ericsson 2003).

There also exists a vast array of mental techniques alleged to boost various skills, such as creativity training, speed reading methods (Calef et al. 1999), and mind-maps (Buzan 1982; Farrand et al. 2002). It is unclear how widespread such techniques are, and good data regarding their efficacy is often lacking. Even if a technique improves performance on some task under particular conditions that does not necessarily mean that the technique is practically useful. In order for a technique to significantly benefit a person, it would have to be effectively integrated into her everyday work.

Of the mental training techniques, visualization may have the greatest potential for future development. While new memory arts can be developed the need for them is limited thanks to easily accessible external storage (the main exception may be remembering passwords). Serious studies of the efficiency of other mental techniques may be worthwhile. However, their specificity to particular tasks limits them. Methods of taking advantage of brain reorganization, possibly enhanced through nootropics and/or virtual reality training, appear to have general utility.

INFORMATION TECHNOLOGY

External hardware is of course already used for cognitive enhancement, be it pen and paper or computer software like personal organizers. This section can only scratch at the surface of the vast range of information technologies that have a cognitive enhancement function. There is practically no cognitive area where there does not exist external hardware or software amplification.

Many common pieces of software act as cognition-enhancing environments where the software helps give an overview, keep multiple items in memory, and perform routine tasks. Data mining and information visualization tools help produce overview and understanding where the perceptual system cannot handle the amount of data, while specialized tools like expert systems, symbolic math programs, decision support tools, and search agents expand specific skills and capacities.

What is new is the growing interest in creating intimate links between the external systems and the human user through better interaction. The software becomes less an external tool and more of a mediating “exoself.” This can be achieved through mediation, embedding the human within an augmenting “shell,” such as wearable computers (Mann 2001; Mann and Niedzviecki 2001) or virtual reality, or through smart environments, where capabilities of objects in the environment are extended. An example is the ubiquitous computing vision, in which objects would be equipped with unique identities and given ability to communicate with and to support the user (Weiser 1991). A well- designed environment can enhance proactive memory (Sellen et al. 1996) by deliberately bringing previous intentions to mind in the right context.

Another form of memory-enhancing exoself software is remembrance agents (Rhodes and Starner 1996), agents that act as a vastly extended associative memory. The agents have access to a database of previous information, such as a user’s files, e-mail correspondence, etc., and suggest relevant documents based on the current context. Other exoself applications include additions to vision (Mann 1997), team coordination (Fan et al. 2005a, 2005b), face recognition (Singletary and Starner 2000), mechanical prediction (Jebara et al. 1997), and recording emotionally significant events (Healey and Picard 1998).

Given the availability of external memory support, from writing to wearable computers, it appears likely that the crucial form of memory demand will be the ability to link together information into usable concepts and associations rather than storage and retrieval of raw data. Storage and retrieval functions can be offloaded to a great extent from the brain, while the knowledge, strategies, and associations linking the data to skilled cognition so far cannot generally be offloaded.

Wearable computers and personal digital assistants (PDAs) are already intimate devices worn on the body, but there have been proposals for even tighter interfaces. Control of external devices through brain activity has been studied with some success for the last 40 years, although it remains a slow form of signaling (Wolpaw et al. 2000).

The most dramatic potential internal hardware enhancements are brain– computer interfaces. At present development is rapid both on the hardware side, where multielectrode recordings from more than 300 electrodes permanently implanted in the brain are currently state of the art, and on the software side, with computers learning to interpret the signals and commands (Carmena et al. 2003; Nicolelis et al. 2003; Shenoy et al. 2003). Early experiments on humans have shown that it is possible for profoundly paralyzed patients to control a computer cursor using just a single electrode (Kennedy and Bakay 1998) implanted in the brain and a 96 electrode prototype has been demonstrated (Hochberg et al. 2006). Prefrontal recordings enable choice selection with a bandwidth of 6.5 bits/s (Santhanam et al. 2006). Experiments in localized chemical release from implanted chips also suggest the possibility to use neural growth factors to promote patterned local growth and interfacing (Peterman et al. 2004).

Cochlear implants are already widely used, and there is ongoing research in artificial retinas (Alteheld et al. 2004) and functional electric stimulation for paralysis treatment (von Wild et al. 2002). These implants are mainly intended to ameliorate functional deficits and will hardly be attractive for healthy people in the foreseeable future. But the digital parts of the implant can in principle be connected to nearly any kind of software and external hardware (Hochberg et al. 2006). This would enable enhancing uses, such as access to software help, Internet, and virtual reality applications. It has been demonstrated that a healthy volunteer could control a robotic arm using tactile feedback, both in direct adjacency and remotely, as well as a wheelchair and perform simple neural communication with another implant (Warwick et al. 2003). Nondisabled people, however, would most likely achieve the same benefits through eyes, finger, and voice control. Neural implants are unlikely to become common enhancements until a “killer application” that cannot be achieved using external technology is found.

COLLECTIVE ENHANCEMENT

Much of human cognition is distributed across many minds and can be enhanced by developing more efficient forms of collaboration. Cooperative groups can detect deception better than individuals (Frank et al. 2004) and solve many problems better than equal numbers of individuals or even the best individuals (Laughlin et al. 2002; Kerr and Tindale 2004; Laughlin et al. 2006).

In general, the total ability of a group to perform a task increases with the size of the group as long as the members do not need to interact much. If they need to coordinate, the efficiency starts to drop as time has to be spent on coordination rather than work. In a densely connected group this eventually produces a situation where adding people reduces total performance. Reducing the density of the network by adding a hierarchy enables larger groups at the price of information bottlenecks. Social cognitive enhancement would act by either increasing the performance of individual group members (improving overall performance), improve their ability to coordinate (enabling larger groups), or improve the synergies generated by having multiple competencies.

This is an area ideally suited for embedding technologies that mediate group interactions. Virtual workspaces can enable improved pattern recognition (Hayne et al. 2003) and various forms of groupware attempts to facilitate collaboration. However, the greatest enhancements occur when very large groups can be facilitated: the World Wide Web and e-mail are among the most powerful kinds of cognitive enhancement software developed to date. Through the use of such social software, the distributed intelligence of large groups can be shared and harnessed for particular purposes (Surowiecki 2004).

Connected systems allow many people to collaborate in the construction of shared knowledge and solutions: the more individuals that connect, the more powerful the system becomes (Drexler 1991). The information is not just stored in the documents themselves but in their interrelations. When such interconnected information resources exist, automated systems, such as search engines (Kleinberg 1999), can extract a wealth of useful information from them.

Lowered coordination costs enable larger groups to work on common projects. Such groups of shared interests, such as amateur journalist “bloggers” and open source programmers, have demonstrated that they can successfully complete large projects, such as online political campaigns (Drezner and Farrell 2004), the Wikipedia encyclopedia, and the Linux operating system. Systems for online collaboration can incorporate efficient error correction (Raymond 2001; Giles 2005), enabling incremental improvement of product quality over time.

An interesting variant of knowledge aggregation is prediction markets (also known as “information markets” or “idea futures markets”). Here participants trade in predictions of future events, and the prices of these bets tend to reflect the best information available on the probability of whether the events will occur (Hanson et al. 2003). Such markets appear to be self-correcting and resilient, and have been shown to outperform alternative methods of generating probabilistic forecasts, such as opinion polls and expert panels (Hanson et al. 2006).

Social cognitive enhancement represents a convergence of not only information and cognitive technology, but sociology, management, and epistemology. In order to be successful a wide variety of factors must come together, making deliberate design hard. It may not be a coincidence that the most successful systems have been the most open, enabling many different groups to experiment and discover whether they can get it to work for their goals. We seldom notice the vast number of failed attempts because they are overshadowed by the explosive growth of successful systems.

NANOTECHNOLOGY

Nanotechnology has so far not been applied to cognitive enhancement, which is unsurprising given its early state. However, in basic neuroscience research many nanotechnology applications are in use or close to use. Fluorescent nanodots are used in neuroscience research, where they enable direct observation of biomolecule interaction (Mitchell 2001; Weng and Ren 2006). Nanostructured scaffolds are explored in tissue engineering (Silva et al. 2004) and nerve regrowth (Ellis-Behnke et al. 2006). One near future application with great promise is nanostructured neural interfaces (Cheung et al. 2002). Providing the right surface would both help improve signal quality and reliability. Nanoelectrodes may also be threaded through the capillary system, enabling low-invasive neurointerfacing (Llinas et al. 2005). At the very least nanotechnology is an enabler of neuroscience research relevant for cognitive enhancement.

Another near-term application pursued with much commercial interest is drug delivery through nanostructures (Panyam and Labhasetwar 2003; Sahoo and Labhasetwar 2003) or controlled-release microchips (Santini et al. 1999; Grayson et al. 2003). This would enable precision pharmacology as well as possibly gene therapy.

It appears likely that as nanotechnology matures it will become an integral part of nearly all cognitive enhancement methods, be they light, powerful, and portable Internet interfaces or “smart drugs” that release modulators just when they are needed.

CONVERGING METHODS

As we have seen, many current applications already span disciplinary borders. Cognitive enhancement is based on the unity between the biological brain and the mind, and the unity between different kinds of information processing. Changing biological processes enables changes to the mind (and vice versa). Information processing is the same whether a brain or a computer does it. It hence lends itself well to the vision of converging technology.

Convergence enables many extensions of the current possibilities. As an example, take cortical plasticity. Currently it can be increased by attention, TMS, and drugs, such as amphetamines, in order to improve rehabilitation or learning. All three methods achieve the same goal using different means. Sustaining attention on a task requires motivation and can plausibly be improved using various forms of mental training; it has the benefit of being highly selective but requires significant effort. Current drugs are nonspecific and would increase plasticity in other cortical areas than the desired ones, besides effects on other parts of the brain. TMS is specific to a particular cortical area but requires training close in time and space to the stimulation equipment, and the task may be distributed over a large number of cortical areas.

Convergence easily suggests multiple ways these techniques can be improved. Improving TMS in terms of location specificity may be achieved by embedding micro-or nanoparticles close to the area to be modified (this may be particularly suitable for rehabilitation after neurosurgery) that augment the signal or help target it. By placing drugs within the particles they might promote nerve regrowth or dendritic sprouting, possibly triggered by external signals (Sershen and West 2002). The use of magnetic particles has already been explored to concentrate drugs to cancer tumors (Lubbe et al. 1996; Lubbe et al. 2001; Kim et al. 2006): the same mechanism could enable concentrating plasticity increasing drugs to the right cortical region even without surgery. More advanced particles or controlled-release microchips may able to sense local neuromodulator concentrations and regulate their drug release to amplify the selective effect of attention during training, making sure only the areas relevant to the training get affected.

On the macroscale, better sensor systems would enable improved understanding of individual brain chemistry, a prerequisite for finding the optimal combination of enhancer drugs. Wearable computing and other personal sensory devices enable the monitoring not only of body state but also behavior. This could enable personal data mining to find the individual optimum of, for example, blood glucose for different cognitive tasks by comparing it with monitored performance, and then based on measured glucose response to different meals suggest food intake that fits future scheduled demands. By pooling such individual data it would also be possible to make inferences on the general utility of different enhancer methods and the interactions between different factors. Rather than being a top-down academic study it might emerge as voluntary data sharing among users. This would represent an entirely new kind of epidemiological study. The difference is similar to the difference between the web-indices that seek to organize web sites into a predetermined subject hierarchy/ontology, and the current Web 2.0 experiments with generating “folksonomies” from locally tagged data (Shirky 2005). The challenges this form of “folk experimentation” poses in terms of data mining, scientific rigor (even if useful patterns are found the uncontrolled nature of the data may make strict interpretation hard), integrity concerns, and ownership concerns (are participants reimbursed for profitable discoveries?) are obviously great and may provide a very fruitful areas of research.

These scenarios are of course merely extrapolations at present, but demonstrate the synergistic potential of many current enhancement techniques.

DISCUSSION

Cognitive enhancement is already in widespread use, but not recognized as such. The morning coffee, the crossword, the e-mail program, and the cellphone are all part of our cognitive enhancement infrastructure. The new kinds of enhancement discussed in this article may appear unusual, futuristic, or problematic but will likely in time become as prosaic and accepted as the others.

It is easier to improve specialized abilities than general cognition. But the rewards are far greater for general cognition. It comes into play all the time, supporting many tasks—including uses we may not have thought of enhancing. Better memory may help education but it may also help remembering one’s holiday memories and avoid forgetting keys. The overall societal impact of even a small increase in general cognitive function would likely be sizeable and desirable. Economic models of the loss caused by small intelligence decrements due to lead in drinking water predict significant effects of even a few points decrease (Salkever 1995; Muir and Zegarac 2001). Because the models are roughly linear for small changes, they can be inverted to estimate societal effects of improved cognition. The Salkever model estimates the increase in income due to one more IQ point to be 2.1% for men and 3.6% for women. (Herrnstein and Murray 1994) estimate that a 3% increase in overall IQ would reduce the poverty rate by 25%, males in jail by 25%, high-school dropouts by 28%, parentless children by 20%, welfare recipients by 18%, and out-of-wedlock births by 25%.

Cognitive enhancement raises many ethical and social issues but also many practical challenges. Enhancements do have a price. In some cases it is a monetary price tag, but often it is a tradeoff between different abilities. Keeping awake using stimulants prevents the memory consolidation that would have taken place during sleep, and enhanced concentration ability may impair the ability to notice things in peripheral awareness. In some cases these tradeoffs can be predicted in terms of known biology or the evolutionary past of humans (Bostrom and Sandberg 2006), but often we will have to do an empirically based evaluation of what we individually value in a particular situation.

A major concern for all forms of enhancement is risk, both from enhancement itself and its effects (as well as its development in clinical trials). Enhancement users must decide when the benefits outweigh the potential risk, and how to estimate this on the basis of available information, personal goals, and their ways of life. These risks cannot always be accurately determined beforehand, nor may a user be able to defer to experts to judge whether the benefits are, to her, worth the risks. This poses a challenge to many current risk frameworks that are based on reducing the risk for the population at large: enhancement may be so individual and variable that it does not fit into a paternalistic framework. This challenge is further complicated because of the convergent nature of enhancement, which will bring different fields with conflicting risk concepts (e.g., medicine, education, and computing) into overlap. Developing a consistent, technology-independent risk management framework for converging technologies is an important task for the future, necessary for the eventual acceptance of general enhancement.

The reliability of research is also an issue. Many of the cognition-enhancing interventions show small effect sizes, which may necessitate very large epidemiological studies possibly exposing large groups to unforeseen risks.

One of the greatest challenges to developing effective cognitive enhancement is the current research model. Enhancers are tested within a laboratory setting for particular tasks. While this enables exact measurement and elimination of confounders, it does not test whether the enhancers aid real-life tasks and lifestyles. Ecological testing in real-life situations would be more relevant, but is far more expensive, time consuming, and hard to interpret. The “folk experimentation” scenario mentioned above might solve the first two problems but would likely worsen the third.

An interesting exception is military enhancement research, where studies in a more realistic (if still somewhat limited) setting are sometimes pursued. Civilian spin-offs from the current programs are likely, although the research ethics issues of military biomedical research are clearly nontrivial (cf. Pearn 2000; McManus et al. 2005 for a discussion of issues of captive subjects and informed consent).

A major challenge in developing human enhancement technologies is the need for interdisciplinary understanding. The problems facing “neuroeducation” and groupware have already been mentioned. While narrowly focused technical work is necessary, it may be that some of the most fruitful approaches will consist in creatively combining and applying work from multiple disciplines. Seeing cognitive enhancement as one field and as a general goal, rather than as multitude of unrelated pursuits, may enable us to spot many promising research questions and enhancement opportunities that would otherwise be overlooked.