hallucination problem

6 results back to index


pages: 444 words: 117,770

The Coming Wave: Technology, Power, and the Twenty-First Century's Greatest Dilemma by Mustafa Suleyman

"World Economic Forum" Davos, 23andMe, 3D printing, active measures, Ada Lovelace, additive manufacturing, agricultural Revolution, AI winter, air gap, Airbnb, Alan Greenspan, algorithmic bias, Alignment Problem, AlphaGo, Alvin Toffler, Amazon Web Services, Anthropocene, artificial general intelligence, Asilomar, Asilomar Conference on Recombinant DNA, ASML, autonomous vehicles, backpropagation, barriers to entry, basic income, benefit corporation, Big Tech, biodiversity loss, bioinformatics, Bletchley Park, Blitzscaling, Boston Dynamics, business process, business process outsourcing, call centre, Capital in the Twenty-First Century by Thomas Piketty, ChatGPT, choice architecture, circular economy, classic study, clean tech, cloud computing, commoditize, computer vision, coronavirus, corporate governance, correlation does not imply causation, COVID-19, creative destruction, CRISPR, critical race theory, crowdsourcing, cryptocurrency, cuban missile crisis, data science, decarbonisation, deep learning, deepfake, DeepMind, deindustrialization, dematerialisation, Demis Hassabis, disinformation, drone strike, drop ship, dual-use technology, Easter island, Edward Snowden, effective altruism, energy transition, epigenetics, Erik Brynjolfsson, Ernest Rutherford, Extinction Rebellion, facts on the ground, failed state, Fairchild Semiconductor, fear of failure, flying shuttle, Ford Model T, future of work, general purpose technology, Geoffrey Hinton, global pandemic, GPT-3, GPT-4, hallucination problem, hive mind, hype cycle, Intergovernmental Panel on Climate Change (IPCC), Internet Archive, Internet of things, invention of the wheel, job automation, John Maynard Keynes: technological unemployment, John von Neumann, Joi Ito, Joseph Schumpeter, Kickstarter, lab leak, large language model, Law of Accelerating Returns, Lewis Mumford, license plate recognition, lockdown, machine readable, Marc Andreessen, meta-analysis, microcredit, move 37, Mustafa Suleyman, mutually assured destruction, new economy, Nick Bostrom, Nikolai Kondratiev, off grid, OpenAI, paperclip maximiser, personalized medicine, Peter Thiel, planetary scale, plutocrats, precautionary principle, profit motive, prompt engineering, QAnon, quantum entanglement, ransomware, Ray Kurzweil, Recombinant DNA, Richard Feynman, Robert Gordon, Ronald Reagan, Sam Altman, Sand Hill Road, satellite internet, Silicon Valley, smart cities, South China Sea, space junk, SpaceX Starlink, stealth mode startup, stem cell, Stephen Fry, Steven Levy, strong AI, synthetic biology, tacit knowledge, tail risk, techlash, techno-determinism, technoutopianism, Ted Kaczynski, the long tail, The Rise and Fall of American Growth, Thomas Malthus, TikTok, TSMC, Turing test, Tyler Cowen, Tyler Cowen: Great Stagnation, universal basic income, uranium enrichment, warehouse robotics, William MacAskill, working-age population, world market for maybe five computers, zero day

In AI, technical safety also means sandboxes and secure simulations to create provably secure air gaps so that advanced AIs can be rigorously tested before they are given access to the real world. It means much more work on uncertainty, a major focus right now—that is, how does an AI communicate when it might be wrong? One of the issues with LLMs is that they still suffer from the hallucination problem, whereby they often confidently claim wildly wrong information as accurate. This is doubly dangerous given they often are right, to an expert level. As a user, it’s all too easy to be lulled into a false sense of security and assume anything coming out of the system is true.

Brian, 56 artificial capable intelligence (ACI), vii, 77–78, 115, 164, 210 artificial general intelligence (AGI) catastrophe scenarios and, 209, 210 chatbots and, 114 DeepMind founding and, 8 defined, vii, 51 gorilla problem and, 115–16 gradual nature of, 75 superintelligence and, 75, 77, 78, 115 yet to come, 73–74 artificial intelligence (AI) aspirations for, 7–8 autonomy and, 114, 115 as basis of coming wave, 55 benefits of, 10–11 catastrophe scenarios and, 208, 209–11 chatbots, 64, 68, 70, 113–14 Chinese development of, 120–21 choke points in, 251 climate change and, 139 consciousness and, 74, 75 contradictions and, 202 costs of, 64, 68 current applications, 61–62 current capabilities of, 8–9 cyberattacks and, 162–63, 166–67 defined, vii early experiments in, 51–54 efficiency of, 68–69 ego and, 140 ethics and, 254 explanation and, 243 future of, 78 future ubiquity of, 284–85 global reach of, 9–10 hallucination problem and, 243 human brain as fixed target, 67–68 hyper-evolution and, 109 invisibility of, 73 limitations of, 73 medical applications, 110 military applications, 104, 165 Modern Turing Test, 76–77, 78, 115, 190, 210 narrow nature of, 73–74 near-term capabilities, 77 omni-use technology and, 111, 130 openness imperative and, 128–29 potential of, 56, 70, 135 as priority, 60 profit motive and, 134, 135, 136 proliferation of, 68–69 protein structure and, 88–89 red teaming and, 246 regulation attempts, 229, 260–61 research unpredictability and, 130 robotics and, 95, 96, 98 safety and, 241, 243–44 scaling hypothesis, 67–68, 74 self-critical culture and, 270 sentience claims, 72, 75 skepticism about, 72, 179 surveillance and, 193–94, 195, 196 synthetic biology and, 89–90, 109 technological unemployment and, 177–81 Turing test, 75 See also coming wave; deep learning; machine learning arXiv, 129 Asilomar principles, 269–70, 272–73 ASML, 251 asymmetrical impact, 105–7, 234 Atlantis, 5 Atmanirbhar Bharat program (India), 125–26 attention, 63 attention maps, 63 audits, 245–48, 267 Aum Shinrikyo, 212–13, 214 authoritarianism, 153, 158–59, 191–96, 216–17 autocomplete, 63 automated drug discovery, 110 automation, 177–81 autonomy, 105, 113–15, 166, 234 Autor, David, 179 al-Awlaki, Anwar, 171 B backpropagation, 59 bad actor empowerment, 165–66, 208, 266 See also terrorism B corps, 258 Bell, Alexander Graham, 31 Benz, Carl, 24, 285 Berg, Paul, 269–70 BGI Group, 122 bias, 69–70, 239–40 Bioforge, 86 Biological Weapons Convention, 241, 263 biotech.

See climate change Go, 53–54, 113, 117–19, 120 Google corporate power of, 187 DeepMind purchase, 60, 255–57 efficiency and, 68 LaMDA and, 71, 72 large language models and, 66 quantum computing and, 97–98, 122 robotics and, 95 on transformers, 64 Google Scholar, 128 Gopher, 68 gorilla problem, 115–16 governments containment and, 258–63 organizational limitations of, 148–50 See also nation-states GPS (Global Positioning System), 110 GPT-2, 64, 70 GPT-3, 64, 68 GPT-4, 64, 113–14 GPUs, 130, 251 grand bargain, defined, viii Great Britain corporations and, 186, 189 surveillance, 193, 195–96 great power competition. See geopolitics Gutenberg, Johannes, 30, 35 H H1N1 flu, 173–74 hallucination problem, 243 Harvard Wyss Institute, 95 Hassabis, Demis, 8 health care. See medical applications Henrich, Joseph, 28 Heritage Foundation, 257 Hershberg, Elliot, 87 Hezbollah, 196–97 Hidalgo, César, 108 hierarchical planning, 76–77 Hinton, Geoffrey, 59, 60, 130 Hiroshima/Nagasaki bombings, 41–42 Hobbes, Thomas, 216 Homo technologicus, 6 Hugging Face, 199 Human Genome Project, 80–81 Huskisson, William, 131 Hutchins, Marcus, 161 hyper-evolution, 105, 107–9 chips and, 32–33, 57, 81, 108 containment and, 250 large language models and, 66, 68 I India, 125–26, 169–70 Industrial Revolution containment attempts, 39, 40, 281–83 openness imperative and, 127 profit motive and, 133, 134 technology waves and, 28–29 inertial confinement, 100 Inflection AI, 66, 68, 243, 244 information dematerialization and, 55–56 DNA as, 79, 87–88 Institute of Electrical and Electronics Engineers, 241 integrated circuit, 32 intelligence action and, 75–76 corporations and, 186–87 economic value of, 136 gorilla problem, 115–16 prediction and, 62 See also artificial intelligence interconnectedness, 28 Intergovernmental Panel on Climate Change, 138–39 internal combustion engine, 24–25, 26, 35–36 International Atomic Energy Agency, 241 international cooperation, 263–67 internet, 33, 107–8, 202 iPhone, 187 Iran, 165 Israel, 165 J James, Kay Coles, 257 Japan, containment attempts, 39, 40 jobs, technology impact on, 177–81, 261, 262 Joint European Torus, 100 K Kasparov, Garry, 53 Kay, John, 39 Ke Jie, 118–19, 121 Kennan, George F., 37 Keynes, John Maynard, 178 Khan, A.


pages: 688 words: 147,571

Robot Rules: Regulating Artificial Intelligence by Jacob Turner

"World Economic Forum" Davos, Ada Lovelace, Affordable Care Act / Obamacare, AI winter, algorithmic bias, algorithmic trading, AlphaGo, artificial general intelligence, Asilomar, Asilomar Conference on Recombinant DNA, autonomous vehicles, backpropagation, Basel III, bitcoin, Black Monday: stock market crash in 1987, blockchain, brain emulation, Brexit referendum, Cambridge Analytica, Charles Babbage, Clapham omnibus, cognitive dissonance, Computing Machinery and Intelligence, corporate governance, corporate social responsibility, correlation does not imply causation, crowdsourcing, data science, deep learning, DeepMind, Demis Hassabis, distributed ledger, don't be evil, Donald Trump, driverless car, easy for humans, difficult for computers, effective altruism, Elon Musk, financial exclusion, financial innovation, friendly fire, future of work, hallucination problem, hive mind, Internet of things, iterative process, job automation, John Markoff, John von Neumann, Loebner Prize, machine readable, machine translation, medical malpractice, Nate Silver, natural language processing, Nick Bostrom, Northpointe / Correctional Offender Management Profiling for Alternative Sanctions, nudge unit, obamacare, off grid, OpenAI, paperclip maximiser, pattern recognition, Peace of Westphalia, Philippa Foot, race to the bottom, Ray Kurzweil, Recombinant DNA, Rodney Brooks, self-driving car, Silicon Valley, Stanislav Petrov, Stephen Hawking, Steve Wozniak, strong AI, technological singularity, Tesla Model S, The Coming Technological Singularity, The Future of Employment, The Signal and the Noise by Nate Silver, trolley problem, Turing test, Vernor Vinge

Hart, Punishment and Responsibility: Essays in the Philosophy of Law (Oxford: Clarendon Press, 1978). 130Carlsmith and Darley, “Psychological Aspects of Retributive Justice”, in Advances in Experimental Social Psychology, edited by Mark Zanna (San Diego, CA: Elsevier, 2008). 131In evidence to the Royal Commission on Capital Punishment, Cmd. 8932, para. 53 (1953). 132Exodus 21:24, King James Bible. 133John Danaher, “Robots, Law and the Retribution Gap”, Ethics and Information Technology, Vol. 18, No. 4 (December 2016), 299–309. 134Recent experiments conducted by Zachary Mainen involving the use of the hormone serotonin on biological systems may provide one avenue for future AI to experience emotions in a similar manner to humans. See Matthew Hutson, “Could Artificial Intelligence Get Depressed and Have Hallucinations?”, Science Magazine, 9 April 2018, http://​www.​sciencemag.​org/​news/​2018/​04/​could-artificial-intelligence-get-depressed-and-have-hallucinations, accessed 1 June 2018. 135In a gruesome example of public retribution being exacted against insensate “perpetrators”, in 1661 following the restoration of the English monarchy after the English Civil War and the rebublican Protectorate, three of the already deceased regicides who had participated in the execution of Charles I were disinterred from their graves and tried for treason.


Visual Thinking: The Hidden Gifts of People Who Think in Pictures, Patterns, and Abstractions by Temple Grandin, Ph.D.

2013 Report for America's Infrastructure - American Society of Civil Engineers - 19 March 2013, 3D printing, a long time ago in a galaxy far, far away, air gap, Albert Einstein, American Society of Civil Engineers: Report Card, Apollo 11, Apple II, ASML, Asperger Syndrome, autism spectrum disorder, autonomous vehicles, Black Lives Matter, Boeing 737 MAX, Captain Sullenberger Hudson, clean water, cloud computing, computer vision, Computing Machinery and Intelligence, coronavirus, cotton gin, COVID-19, defense in depth, Drosophila, Elon Musk, en.wikipedia.org, GPT-3, Gregor Mendel, Greta Thunberg, hallucination problem, helicopter parent, income inequality, industrial robot, invention of movable type, Isaac Newton, James Webb Space Telescope, John Nash: game theory, John von Neumann, Jony Ive, language acquisition, longitudinal study, Mark Zuckerberg, Mars Rover, meta-analysis, Neil Armstrong, neurotypical, pattern recognition, Peter Thiel, phenotype, ransomware, replication crisis, Report Card for America’s Infrastructure, Robert X Cringely, Saturday Night Live, self-driving car, seminal paper, Silicon Valley, Skinner box, space junk, stem cell, Stephen Hawking, Steve Jobs, Steve Wozniak, Tacoma Narrows Bridge, TaskRabbit, theory of mind, TikTok, twin studies, unpaid internship, upwardly mobile, US Airways Flight 1549, warehouse automation, warehouse robotics, web application, William Langewiesche, Y Combinator

The odds of something offensive coming out is 100 percent.” AI applications are being developed for simulations and analytics, and in industry, transportation, cybersecurity, and the military. What are the failsafes? Would you want an AI program running a nuclear reactor? What if the AI operator started hallucinating because a hacker inserted a feedback loop that forced it to perceive the high pressures and temperatures of a meltdown that did not exist? Maybe it would create an actual meltdown. Some computer scientists will admit that they are not completely sure how AI works. In an article by Arthur I.


pages: 584 words: 170,388

pages: 574 words: 164,509

Superintelligence: Paths, Dangers, Strategies by Nick Bostrom

agricultural Revolution, AI winter, Albert Einstein, algorithmic trading, anthropic principle, Anthropocene, anti-communist, artificial general intelligence, autism spectrum disorder, autonomous vehicles, backpropagation, barriers to entry, Bayesian statistics, bioinformatics, brain emulation, cloud computing, combinatorial explosion, computer vision, Computing Machinery and Intelligence, cosmological constant, dark matter, DARPA: Urban Challenge, data acquisition, delayed gratification, Demis Hassabis, demographic transition, different worldview, Donald Knuth, Douglas Hofstadter, driverless car, Drosophila, Elon Musk, en.wikipedia.org, endogenous growth, epigenetics, fear of failure, Flash crash, Flynn Effect, friendly AI, general purpose technology, Geoffrey Hinton, Gödel, Escher, Bach, hallucination problem, Hans Moravec, income inequality, industrial robot, informal economy, information retrieval, interchangeable parts, iterative process, job automation, John Markoff, John von Neumann, knowledge worker, Large Hadron Collider, longitudinal study, machine translation, megaproject, Menlo Park, meta-analysis, mutually assured destruction, Nash equilibrium, Netflix Prize, new economy, Nick Bostrom, Norbert Wiener, NP-complete, nuclear winter, operational security, optical character recognition, paperclip maximiser, pattern recognition, performance metric, phenotype, prediction markets, price stability, principal–agent problem, race to the bottom, random walk, Ray Kurzweil, recommendation engine, reversible computing, search costs, social graph, speech recognition, Stanislav Petrov, statistical model, stem cell, Stephen Hawking, Strategic Defense Initiative, strong AI, superintelligent machines, supervolcano, synthetic biology, technological singularity, technoutopianism, The Coming Technological Singularity, The Nature of the Firm, Thomas Kuhn: the structure of scientific revolutions, time dilation, Tragedy of the Commons, transaction costs, trolley problem, Turing machine, Vernor Vinge, WarGames: Global Thermonuclear War, Watson beat the top human players on Jeopardy!, World Values Survey, zero-sum game

It could inspect each one, over and over, to reduce the risk that any of the paperclips fail to meet the design specifications. It could build an unlimited amount of computronium in an effort to clarify its thinking, in the hope of reducing the risk that it has overlooked some obscure way in which it might have somehow failed to achieve its goal. Since the AI may always assign a nonzero probability to having merely hallucinated making the million paperclips, or to having false memories, it would quite possibly always assign a higher expected utility to continued action—and continued infrastructure production—than to halting. The claim here is not that there is no possible way to avoid this failure mode.


pages: 1,028 words: 267,392