all AI news
QuACK: Accelerating Gradient-Based Quantum Optimization with Koopman Operator Learning
May 7, 2024, 4:45 a.m. | Di Luo, Jiayu Shen, Rumen Dangovski, Marin Solja\v{c}i\'c
cs.LG updates on arXiv.org arxiv.org
Abstract: Quantum optimization, a key application of quantum computing, has traditionally been stymied by the linearly increasing complexity of gradient calculations with an increasing number of parameters. This work bridges the gap between Koopman operator theory, which has found utility in applications because it allows for a linear representation of nonlinear dynamical systems, and natural gradient methods in quantum optimization, leading to a significant acceleration of gradient-based quantum optimization. We present Quantum-circuit Alternating Controlled Koopman learning …
abstract application applications arxiv complexity computing cs.ai cs.lg found gap gradient key math.oc optimization parameters physics.comp-ph quant-ph quantum quantum computing theory type utility work
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Software Engineer for AI Training Data (School Specific)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Python)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Tier 2)
@ G2i Inc | Remote
Data Engineer
@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US