Abstract
Modern digital electronics support remarkably reliable computing, especially given the challenge of controlling nanoscale logical components that interact in fluctuating environments. However, we demonstrate that the high-reliability limit is subject to a fundamental error–energy-efficiency tradeoff that arises from time-symmetric control: requiring a low probability of error causes energy consumption to diverge as the logarithm of the inverse error rate for nonreciprocal logical transitions. The reciprocity (self-invertibility) of a computation is a stricter condition for thermodynamic efficiency than logical reversibility (invertibility), the latter being the root of Landauer's work bound on erasing information. Beyond engineered computation, the results identify a generic error–dissipation tradeoff in steady-state transformations of genetic information carried out by biological organisms. The lesson is that computational dissipation under time-symmetric control cannot reach, and is often far above, the Landauer limit. In this way, time-asymmetry becomes a design principle for thermodynamically efficient computing.
- Received 14 September 2019
- Accepted 24 August 2020
DOI:https://doi.org/10.1103/PhysRevResearch.2.033524
Published by the American Physical Society under the terms of the Creative Commons Attribution 4.0 International license. Further distribution of this work must maintain attribution to the author(s) and the published article's title, journal citation, and DOI.
Published by the American Physical Society