Quantum computers are getting bigger, but there are still few practical ways to take advantage of their extra computing power. To get over this hurdle, researchers are designing algorithms to ease the transition from classical to quantum computers. In a new study in Nature, researchers unveil an algorithm that reduces the statistical errors, or noise, produced by quantum bits, or qubits, in crunching chemistry equations.
Developed by Columbia chemistry professor David Reichman and postdoc Joonho Lee with researchers at Google Quantum AI, the algorithm uses up to 16 qubits on Sycamore, Google's 53-qubit computer, to calculate ground state energy, the lowest energy state of a molecule. "These are the largest quantum chemistry calculations that have ever been done on a real quantum device," Reichman said.
To read more, click here.