Share this article:
How Quantum Mechanics Can Change Computing
In early July, Google announced that it will expand its commercially available cloud computing services to include quantum computing. A similar service has been available from IBM since May. These aren’t services most regular people will have a lot of reason to use yet. But making quantum computers more accessible will help government, academic and corporate research groups around the world continue their study of the capabilities of quantum computing.
Understanding how these systems work requires exploring a different area of physics than most people are familiar with. From everyday experience we are familiar with what physicists call “classical mechanics,” which governs most of the world we can see with our own eyes, such as what happens when a car hits a building, what path a ball takes when it’s thrown and why it’s hard to drag a cooler across a sandy beach.
Quantum mechanics, however, describes the subatomic realm – the behavior of protons, electrons and photons. The laws of quantum mechanics are very different from those of classical mechanics and can lead to some unexpected and counterintuitive results, such as the idea that an object can have negative mass.
Physicists around the world – in government, academic and corporate research groups – continue to explore real-world deployments of technologies based on quantum mechanics. And computer scientists, including me, are looking to understand how these technologies can be used to advance computing and cryptography.
A brief introduction to quantum physics
In our regular lives, we are used to things existing in a well-defined state: A light bulb is either on or off, for example. But in the quantum world, objects can exist in a what is called a superposition of states: A hypothetical atomic-level light bulb could simultaneously be both on and off. This strange feature has important ramifications for computing.
The smallest unit of information in classical mechanics – and, therefore, classical computers – is the bit, which can hold a value of either 0 or 1, but never both at the same time. As a result, each bit can hold just one piece of information. Such bits, which can be represented as electrical impulses, changes in magnetic fields, or even a physical on-off switch, form the basis for all calculation, storage and communication in today’s computers and information networks.
Qubits – quantum bits – are the quantum equivalent of classical bits. One fundamental difference is that, due to superposition, qubits can simultaneously hold values of both 0 and 1. Physical realizations of qubits must inherently be at an atomic scale: for example, in the spin of an electron or the polarization of a photon.
An explanation of quantum mechanics, in terms of how well you remember someone’s name when you see him.
Computing with qubits
Another difference is that classical bits can be operated on independently of each other: Flipping a bit in one location has no effect on bits in other locations. Qubits, however, can be set up using a quantum-mechanical property called entanglement so that they are dependent on each other – even when they are far apart. This means that operations performed on one qubit by a quantum computer can affect multiple other qubits simultaneously. This property – akin to, but not the same as, parallel processing – can make quantum computation much faster than in classical systems.
Large-scale quantum computers – that is, quantum computers with hundreds of qubits – do not yet exist, and are challenging to build because they require operations and measurements to be done on a atomic scale. IBM’s quantum computer, for example, currently has 16 qubits, and Google is promising a 49-qubit quantum computer – which would be an astounding advance – by the end of the year. (In contrast, laptops currently have multiple gigabytes of RAM, with a gigabyte being eight billion classical bits.)
A physics professor untangles entanglement.
A powerful tool
Notwithstanding the difficulty of building working quantum computers, theorists continue to explore their potential. In 1994, Peter Shor showed that quantum computers could quickly solve the complicated math problems that underlie all commonly used public-key cryptography systems, like the ones that provide secure connections for web browsers. A large-scale quantum computer would completely compromise the security of the internet as we know it. Cryptographers are actively exploring new public-key approaches that would be “quantum-resistant,” at least as far as they currently know.
Interestingly, the laws of quantum mechanics can also be used to design cryptosystems that are, in some senses, more secure than their classical analogs. For example, quantum key distribution allows two parties to share a secret no eavesdropper can recover using either classical or quantum computers. Those systems – and others based on quantum computers – may become useful in the future, either widely or in more niche applications. But a key challenge is getting them working in the real world, and over large distances.
This article was originally published on The Conversation website under a Creative Commons Attribution 4.0 International License. Read the original article here.
If you found this article interesting, you may be interested in browsing related content on ScienceDirect. We are pleased to offer you a free chapter from the Quantum Inspired Computational Intelligence book called “Quantum computing and supervised machine learning: Training, model selection, and error estimation.”
Visit elsevier.com to access computer science and more! Use discount code STC317 at checkout and save up to 30% on your very own copy!
Computing functionality is ubiquitous. Today this logic is built into almost any machine you can think of, from home electronics and appliances to motor vehicles, and it governs the infrastructures we depend on daily — telecommunication, public utilities, transportation. Maintaining it all and driving it forward are professionals and researchers in computer science, across disciplines including:
- Computer Architecture and Computer Organization and Design
- Data Management, Big Data, Data Warehousing, Data Mining, and Business Intelligence (BI)
- Human Computer Interaction (HCI), User Experience (UX), User Interface (UI), Interaction Design and Usability
- Artificial intelligence (AI)