Imagine developing a drug in a fraction of the usual time. At a fraction of the cost. Or storing the contents of Wikipedia in a jar, in a fridge, with inexpensive, organic materials. Or, for that matter, using only a tiny proportion of the energy currently devoured by data processing. Future technologies like quantum computing, DNA storage and all-optical computing promise all this and more – much sooner than you might think. Accompanying these exotic approaches are exciting advances in silicon technologies, such as novel architectures, which will arrive in the very near term.

As revolution ferments, what can you do to make sure your business seizes the future of computing and is not a casualty of it? The answer is to take it one step at a time. Gain a working knowledge of the technologies, get a feel for how they can help, and then work out your first steps.

So, on that note, let’s get down to business… 

‘Forty-two!’ yelled Loonquawl. ‘Is that all you've got to show for seven and a half million years' work?’ ‘I checked it very thoroughly,’ said the computer, ‘and that quite definitely is the answer. I think the problem, to be quite honest with you, is that you've never actually known what the question is.’
Douglas Adams
The Hitchhiker's Guide to the Galaxy

First thing to appreciate is that the new order will almost certainly not replace the classical, silicon-based central processing unit (CPU) computing that we know and use today. Rather they will combine as part of a hybrid system to provide a specific capability – a computing superpower, as it were.

In the expandable sections below, we dissect the transformative computing technologies that will land in the next two to 10 years. We reveal how they work, the special capabilities they offer and what you can do to embrace them now. But let’s be clear. The best place for you to start is with the future needs, aspirations and crucial challenges of your business. Then you’ll know what the question is, and you can plan the best computing technology pathway to the future you desire. After all, no-one wants the answer to be ‘42’.

  • It’s tough to find a word that adequately describes the impact of computers on how we work, communicate, learn, and travel. The rate of progress has been equally staggering – there is more compute power in our smartphones than in the rocket that sent Neil Armstrong to the Moon in 1969, all enabled by incredible innovation in the design and fabrication of silicon transistor chips, the bedrock of classical compute.

    But this rate of progress – the doubling of processing speed or power every two years or so - is in jeopardy . These improvements are driven by reducing the size of the chip features, but achieving ever smaller geometries is increasingly challenging, expensive, and time-consuming.

    Of course, not all products require the general-purpose performance capabilities of a CPU chip, and limitations of CPU chips can be addressed by using a custom chip that optimises a particular performance metric. AI chips that have been designed to achieve high-speed AI processing are one example of this. However, even for these chips, it can take months just to get a prototype from a foundry, slowing your innovation and increasing the time to market. And if you need only low volumes, the high initial development cost might also mean increased cost per unit.

    And lastly there are some problems that are simply intractable with classical digital compute based on bits and bytes. Add to all this the increasing concerns about the sustainability of computing applications as the world creates, processes and stores ever more data and it is obvious that we need a new approach to computing.

    The good news is that all these factors are driving innovation in numerous new computing technologies. The not-so-good news is that none of them will provide a one-size-fits-all solution, so it will be critical for businesses to understand exactly what they need.
  • Given the ubiquity, usefulness, and relatively low cost of our existing silicon CPU technologies, it is not surprising that a more effective drop-in replacement has not yet been identified. Newer computing technologies are being developed to fill a particular niche in terms of capability and, as a result, will almost certainly be used as part of a hybrid system that combines them with classical silicon chip computing. Further, these new technologies have different time horizons to maturity (see Figure below).
    Future of computing timeline
    Based on what we’ve seen to date, there are three broad drivers for selecting novel compute technologies and they roughly correlate with time to maturity.

    a) Pushing the bounds of silicon

    There is a strong argument for trying to get more out of silicon, given the decades of research and development to date. Silicon is plentiful and inexpensive compared to other semiconductor materials; we understand its properties and have facilities to fabricate and package devices in large volumes; and we have practical user interfaces, operating systems, and programming languages. Silicon also offers a straightforward route for integration of components, a major driver of research into silicon photonics, for example. Silicon offers some of the nearer-term future compute technologies. Novel Architectures (described in Section 3.5) started to be a reality one or two years ago, while neuromorphic computing and photonic computing, both based on silicon, are likely to be commercially available to the mass market in five or six years.

    b) Achieving a step change in performance, cost or power consumption using a novel approach

    Our computing performance has come a long way since the first transistor – we have seen a trillion-fold increase of computational speed measured in Floating Operations Per Second (FLOPS) over the past 60 years, while the cost of silicon chips has decreased by a factor of 20%-30% annually. So, what do we mean by a step change? Alternative materials systems and approaches could leapfrog our current capabilities with silicon:

    - All-optical photonic computing could offer an increase in processing speed of 10 or 100 times
    - Writing information into DNA offers the ability to store vast amounts of data very inexpensively in a smaller and passive form, thereby reducing energy cost in comparison to traditional silicon.
    - And neuromorphic computers using novel, even quantum, materials could enable much more compact, low power devices that behave more like a human brain

    c) Making the impossible possible

    Despite the great advances in classical computing, some computational problems are still intractable. For example, the Travelling Salesman Problem is a classical problem that tries to find the optimal route between two locations, stopping only once at every intermediate point required. It can be easily solved for a few intermediate points, but the computational time required grows exponentially as more locations need to be considered. This is a daily problem that supply chain and logistic operations face that would currently take a classical computer thousands of years to solve, even without taking into account external factors such as traffic jams, diversions, and delivery times.

    Quantum computing has gained much attention lately because it offers a way to overcome the limitations of digital bits, thereby paving the way for us to simulate and explore complex physical systems, such as large molecules, in a way that is not currently possible even with the most powerful supercomputers. Since simulations can be run in a fraction of the time of physical trials, quantum computing offers the possibility to reduce development time for drugs, for example, or accelerate discovery of new materials.
  • With the physical limitations of reducing the size of foundational electronic components (i.e., transistors and diodes), the future of silicon-based electronics will be dependent on layered designs. So far, these basic components have been manufactured in one-layer wafers, but now multi-layer designs are being considered to keep producing even quicker and smaller chips for a lower price. These computational improvements aim, for instance, to allow personalised chip designs that cope with AI algorithms needs and provide enhanced experience for users.
  • Neuromorphic computing does not follow a sequential processing as traditional computers do, instead they try to imitate how neurons process information. Our neurons are stimulated through spikes that can trigger one or multiple neurons, which can then activate others they are connected with. This means that, like neurons, neuromorphic computing will be capable of parallel and targeted processing for a define set of neurons and interconnections. Its potential will be explored, for instance, by algorithms and sensing systems that try to imitate the human thinking.
  • The vast majority of computing today makes use of the energy contained in electrons to process data. The fundamental idea of photonic computing is to replace the electrons in digital computers with photons, using light waves to process and store data instead. Due to the speed of light being unsurpassable, photonics provides a theoretical minimum latency and therefore makes light a better computing medium compared to electricity. In principle, we could also observe a 10-50x the bandwidth improvement over traditional computing due to the bits travelling at the speed of light. Furthermore, a 10x energy efficiency can be obtained with optical computing as one can increase the processing power without having to increase the electricity supply, which is the case for classical electronic computers.
  • Biological computing uses biologically derived molecules such as DNA and proteins to perform computation, instead of the electrical signals used in today’s digital computers. As a result of advances in nanobiotechnology, researchers have discovered methods of programming living cells to respond in predictable ways to certain chemical inputs. While programming such cells is costly and labour intensive in the first instance, it becomes extremely cost-effective to grow billions more once this is achieved. Biological computing has already been successfully used for cold data storage by encoding information in DNA but will eventually be able to process data with very high reliability. 
  • Quantum computing is based on the physical phenomena where electrons can simultaneously behave as material particles or waves. This property changes the basic idea of storing and processing information in the form of binary bits (0 or 1) and instead using qubits (any number from 0 to 1). Particles will be able to represent any of an infinite number of states and interact with other particles in infinite different ways. Therefore, quantum computers have the potential of solving very complex problems faster than traditional computers.

What does all this mean for you?

The future of computing comprises an array of technologies, all maturing at different times. Many are still evolving, carry a large price tag and present a risk in terms of return. Knowing how and when to invest isn’t easy, so where to start?

Most, if not all, of these technologies are being developed for a specific application or capability; to overcome a particular limitation of existing CPU silicon chips. The objective is not a complete replacement of classical computing. We expect a hybrid approach even for the most exotic of the approaches, like quantum computing.

The way forward is to view the technologies as a way to achieve better capabilities. Work out which capabilities you need to reach your goals, and you’ll be able to fit the new technologies to your needs – in terms of both technical performance and business value. Discover more in the Innovation Briefing that accompanies this page: Future of Computing – the new commercial horizon.

New insights

Discover the developments shaping the future of industry
What are novel silicon architectures?

AI accelerators: Pushing the boundaries of intelligent devices through novel silicon architectures

2023年 10月 6日

Novel silicon architectures offer exciting opportunities for intelligent devices using AI, promising incredible growth for the future of silicon chip design. 

  • AI和数据分析
  • 连接和IoT
  • 电子技术和ASICs
Quantum cloud computing

Quantum cloud computing plus digital service innovation will equal commercial advantage

2023年 03月 1日

What is quantum computing? That’s a question still being asked by much of the world’s C-suite. But here at CC our multidisciplinary teams of physicists, engineers and designers are already helping to plot a credible pathway for a future powered by quantum technologies…

  • 云/可扩展系统
  • 数字化服务
  • 创新管理
  • 战略咨询
2023 technology predictions from Cambridge Consultants

Technology trends and predictions for 2023, with an eye on sustainability

2023年 01月 26日

What are the key technology trends for 2023? A team of Cambridge Consultants innovation experts have taken the sustainability imperative into account. Our team share their thoughts and predictions, partly through the lens of energy and sustainability… 

  • AI和数据分析
  • 连接和IoT
  • 数字化服务
  • 无线电系统
  • 可持续性
Quantum for pharma

Quantum computing for pharma and life sciences – opportunities are on the horizon

2023年 02月 21日

What can quantum computing and quantum technology do for the pharma and life sciences industry? After reviewing the latest signs of quantum innovation in practice across Europe, Elizabeth Thomas has conferred with Cambridge Consultants’ quantum experts and now shares her insights into how organisations can start implementing an action plan… 

  • AI和数据分析
  • 物理科学
  • 战略咨询
  • Synthetic biology

Meet our future of computing experts

The team is ready to connect and talk through your challenges and ambitions

Martin Brock

Chief Technology Officer

Dr Heidi Pinkney

Senior Consultant, Applied Science
Aidong Xu

Dr Aidong Xu

Head of Semiconductor
Bryan Donoghue

Bryan Donoghue

Head of Digital Hardware

Jez Stark

SVP, Client

Dr Sally Epstein

Head of Strategic Technology

Edmund Owen

Principal Quantum Physicist
Sara Llena

Sara Llena

Consultant, Technology Strategy
Georga Liatis

George Liatis

Consultant, Technology Strategy

Talk to one of our experts