来自用户 尹志 的文献。
当前共找到 54 篇文献分享,本页显示第 1 - 20 篇。
1.
尹志 (2026-04-30 21:57):
#paper, Mechanistic insights into ASO-RNA complexation: Advancing antisense oligonucleotide design strategies,DOI:https://doi.org/10.1016/j.omtn.2024.102351, 文章通过md模拟,发现aso与mRNA的结合不仅取决于常规的碱基互补原则,还和 mRNA三维结构比如发夹环的动态可及性和三联碱基对的形成有关。这给未来aso的设计 提出了很有启发的一套思路,毕竟字面上的序列在实际分子层面,还是三维的。 在蛋白质的计算结构生物学开展的如此热烈的背景下,会有更多有价值的工作出现。
2.
尹志 (2026-03-31 23:30):
#paper, Quantum-HPC hybrid computation of biomolecular excited-state energies, DOI: 10.48550/arXiv.2601.15677. 通过ONIOM框架,结合TE-QSCI算法,在离子阱方案上实现了视网膜醛的光异构化的S0、S1以及T0的能量计算。非常好的量子+HPC混合计算的例子。
arXiv, 2026-01-22T05:57:54Z. DOI: 10.48550/arXiv.2601.15677
Kentaro Yamamoto, Riku Masui, Takahito Nakajima, Miwako Tsuji, Mitsuhisa Sato, Peter Schow, Lukas Heidemann, Matthew Burke, Philipp Seitz, Oliver J. Backhouse ... >>>
Kentaro Yamamoto, Riku Masui, Takahito Nakajima, Miwako Tsuji, Mitsuhisa Sato, Peter Schow, Lukas Heidemann, Matthew Burke, Philipp Seitz, Oliver J. Backhouse, Juan W. Pedersen, John Children, Craig Holliman, Nathan Lysne, Daichi Okuno, Seyon Sivarajah, David Muñoz Ramo, Alex Chernoguzov, Ross Duncan <<<
Abstract:
We develop a workflow within the ONIOM framework and demonstrate it on the hybrid computing system consisting of the supercomputer Fugaku and the Quantinuum Reimei trapped-ion quantum computer. This hybrid platform extends the layered approach for biomolecular chemical reactions to accurately treat the active site, such as a protein, and the large and often weakly correlated molecular environment. Our result marks a significant milestone in enabling scalable and accurate simulation of complex biomolecular reactions
3.
尹志 (2026-02-28 23:14):
#paper,DOI: arXiv:2601.10144,Bridging Superconducting and Neutral-Atom Platforms for Efficient Fault-Tolerant Quantum Architectures, 本文提出了一种整合超导和中性原子方案的混合量子计算架构,且面向容错。很有启发性很有前瞻性。考虑到不同量子计算体系的特点,混合方案确实有机会在未来带来有价值的变革。今年我们也会从问题域视角进行混合架构的探索。
arXiv, 2026-01-15T07:39:05Z. DOI: 10.48550/arXiv.2601.10144
Xiang Fang, Jixuan Ruan, Sharanya Prabhu, Ang Li, Travis Humble, Dean Tullsen, Yufei Ding
Abstract:
The transition to the fault-tolerant era exposes the limitations of homogeneous quantum systems, where no single qubit modality simultaneously offers optimal operation speed, connectivity, and scalability. In this work, we propose a strategic approach to Heterogeneous Quantum Architectures (HQA) that synthesizes the distinct advantages of the superconducting (SC) and neutral atom (NA) platforms. We explore two architectural role assignment strategies based on hardware characteristics: (1) We offload the latency-critical Magic State Factory (MSF) to fast SC devices while performing computation… >>>
The transition to the fault-tolerant era exposes the limitations of homogeneous quantum systems, where no single qubit modality simultaneously offers optimal operation speed, connectivity, and scalability. In this work, we propose a strategic approach to Heterogeneous Quantum Architectures (HQA) that synthesizes the distinct advantages of the superconducting (SC) and neutral atom (NA) platforms. We explore two architectural role assignment strategies based on hardware characteristics: (1) We offload the latency-critical Magic State Factory (MSF) to fast SC devices while performing computation on scalable NA arrays, a design we term MagicAcc, which effectively mitigates the resource-preparation bottleneck. (2) We explore a Memory-Compute Separation (MCSep) paradigm that utilizes NA arrays for high-density qLDPC memory storage and SC devices for fast surface-code processing. Our evaluation, based on a comprehensive end-to-end cost model, demonstrates that principled heterogeneity yields significant performance gains. Specifically, our designs achieve $752\times$ speedup over NA-only baselines on average and reduce the physical qubit footprint by over $10\times$ compared to SC-only systems. These results chart a clear pathway for leveraging cross-modality interconnects to optimize the space-time efficiency of future fault-tolerant quantum computers. <<<
4.
尹志 (2026-01-31 23:53):
#paper https://arxiv.org/abs/2601.21571. arxiv 2026. Shaping capabilities with token-level data filtering。文档级过滤过渡到Token 级过滤确实是很直接的想法,但用良好的工程实现获得洞见,确实是alec的风格。
arXiv, 2026-01-29T11:34:01Z. DOI: 10.48550/arXiv.2601.21571
Neil Rathi, Alec Radford
Abstract:
Current approaches to reducing undesired capabilities in language models are largely post hoc, and can thus be easily bypassed by adversaries. A natural alternative is to shape capabilities during pretraining itself. On the proxy task of removing medical capabilities, we show that the simple intervention of filtering pretraining data is highly effective, robust, and inexpensive at scale. Inspired by work on data attribution, we show that filtering tokens is more effective than filtering documents, achieving the same hit to undesired capabilities at a lower cost to benign ones. Training models… >>>
Current approaches to reducing undesired capabilities in language models are largely post hoc, and can thus be easily bypassed by adversaries. A natural alternative is to shape capabilities during pretraining itself. On the proxy task of removing medical capabilities, we show that the simple intervention of filtering pretraining data is highly effective, robust, and inexpensive at scale. Inspired by work on data attribution, we show that filtering tokens is more effective than filtering documents, achieving the same hit to undesired capabilities at a lower cost to benign ones. Training models spanning two orders of magnitude, we then demonstrate that filtering gets more effective with scale: for our largest models, token filtering leads to a 7000x compute slowdown on the forget domain. We also show that models trained with token filtering can still be aligned on the forget domain. Along the way, we introduce a methodology for labeling tokens with sparse autoencoders and distilling cheap, high-quality classifiers. We also demonstrate that filtering can be robust to noisy labels with sufficient pretraining compute. <<<
5.
尹志 (2025-12-31 23:41):
#paper doi: https://doi.org/10.1016/j.future.2024.04.060. Quantum-centric Supercomputing for Materials Science: A Perspective on Challenges and Future Directions. 大综述,讲了量子计算为中心的计算范式,在材料科学中的算法、应用及方向。对多个材料科学的案例进行了讲解,算法部分的综述也很系统。可以说是量子计算for材料科学最优的概览素材之一。甚至对其他类似领域如药物发现等也有很好的借鉴意义。
Yuri Alexeev, Maximilian Amsler, Marco Antonio Barroca, Sanzio Bassini, Torey Battelle, Daan Camps, David Casanova, Young Jay Choi, Frederic T. Chong, Charles Chung ... >>>
Yuri Alexeev, Maximilian Amsler, Marco Antonio Barroca, Sanzio Bassini, Torey Battelle, Daan Camps, David Casanova, Young Jay Choi, Frederic T. Chong, Charles Chung, Christopher Codella, Antonio D. Córcoles, James Cruise, Alberto Di Meglio, Ivan Duran, Thomas Eckl, Sophia Economou, Stephan Eidenbenz, Bruce Elmegreen, Clyde Fare, Ismael Faro, Cristina Sanz Fernández, Rodrigo Neumann Barros Ferreira, Keisuke Fuji, Bryce Fuller, Laura Gagliardi, Giulia Galli, Jennifer R. Glick, Isacco Gobbi, Pranav Gokhale, Salvador de la Puente Gonzalez, Johannes Greiner, Bill Gropp, Michele Grossi, Emanuel Gull, Burns Healy, Matthew R. Hermes, Benchen Huang, Travis S. Humble, Nobuyasu Ito, Artur F. Izmaylov, Ali Javadi-Abhari, Douglas Jennewein, Shantenu Jha, Liang Jiang, Barbara Jones, Wibe Albert de Jong, Petar Jurcevic, William Kirby, Stefan Kister, Masahiro Kitagawa, Joel Klassen, Katherine Klymko, Kwangwon Koh, Masaaki Kondo, Dog̃a Murat Kürkçüog̃lu, Krzysztof Kurowski, Teodoro Laino, Ryan Landfield, Matt Leininger, Vicente Leyton-Ortega, Ang Li, Meifeng Lin, Junyu Liu, Nicolas Lorente, Andre Luckow, Simon Martiel, Francisco Martin-Fernandez, Margaret Martonosi, Claire Marvinney, Arcesio Castaneda Medina, Dirk Merten, Antonio Mezzacapo, Kristel Michielsen, Abhishek Mitra, Tushar Mittal, Kyungsun Moon, Joel Moore, Sarah Mostame, Mario Motta, Young-Hye Na, Yunseong Nam, Prineha Narang, Yu-ya Ohnishi, Daniele Ottaviani, Matthew Otten, Scott Pakin, Vincent R. Pascuzzi, Edwin Pednault, Tomasz Piontek, Jed Pitera, Patrick Rall, Gokul Subramanian Ravi, Niall Robertson, Matteo A.C. Rossi, Piotr Rydlichowski, Hoon Ryu, Georgy Samsonidze, Mitsuhisa Sato, Nishant Saurabh, Vidushi Sharma, Kunal Sharma, Soyoung Shin, George Slessman, Mathias Steiner, Iskandar Sitdikov, In-Saeng Suh, Eric D. Switzer, Wei Tang, Joel Thompson, Synge Todo, Minh C. Tran, Dimitar Trenev, Christian Trott, Huan-Hsin Tseng, Norm M. Tubman, Esin Tureci, David García Valiñas, Sofia Vallecorsa, Christopher Wever, Konrad Wojciechowski, Xiaodi Wu, Shinjae Yoo, Nobuyuki Yoshioka, Victor Wen-zhe Yu, Seiji Yunoki, Sergiy Zhuk, Dmitry Zubarev <<<
6.
尹志 (2025-11-30 19:01):
#paper doi: 10.1101/2025.10.10.681530 Protein Hunter: exploiting structure hallucination within diffusion for protein design。基于生成式AI的蛋白质设计已经成为一种卓有成效的计算范式,但现在大多数的诟病来源于它的准确稳定及泛化能力。我非常认同本文提出的观点,即不需要对生成式AI蛋白质设计“要求过高”,即使存在幻觉,那么幻觉本身就是一种自洽的模式。我们可以通过某些特定的metric去筛选由幻觉产生的结果。考虑到当前已经有很多优秀的模型对蛋白质本身的规律(规律子集)有很好的描述,那么将这样的模型改造为生成器,通过其它手段去进行筛选,可以更充分的利用现有的大量模型。
Yehlin Cho, Griffin Rangel, Gaurav Bhardwaj, Sergey Ovchinnikov
Abstract:
1
Abstract
Interactions between proteins and other biomolecules underlie nearly all biological processes, yet designing such interactions de novo remains challenging. Capturing their specific interactions and co-optimizing sequence and structure are difficult and often require extensive computation. We present Protein Hunter, a fast, fine-tuning-free framework for de novo protein design. Starting from an all-X sequence, we find diffusion-based structure prediction models hallucinate reasonable looking structures that can be further improved through iterative sequence re-design and str… >>>
1<br> Abstract<br> Interactions between proteins and other biomolecules underlie nearly all biological processes, yet designing such interactions de novo remains challenging. Capturing their specific interactions and co-optimizing sequence and structure are difficult and often require extensive computation. We present Protein Hunter, a fast, fine-tuning-free framework for de novo protein design. Starting from an all-X sequence, we find diffusion-based structure prediction models hallucinate reasonable looking structures that can be further improved through iterative sequence re-design and structure re-prediction. This lightweight strategy achieves high AlphaFold3 in silico success rates across both unconditional and conditional generation tasks, including binders to proteins, cyclic peptides, small molecules, DNA, and RNA. Protein Hunter also supports multi-motif scaffolding and partial redesign, providing a general and efficient platform for de novo protein design across diverse molecular targets. <<<
7.
尹志 (2025-10-31 16:37):
#paper Quantum computing and chemistry. doi: 10.1016/j.xcrp.2024.102105 文章从硬件、软件、化学应用(主要是算法)层面综述了现在量子计算在大化学领域的进展,很全面(259篇参考文献)。不过对于纠错算法、错误缓解等算法的讨论比较少。我觉得要想在至少NISQ的时期做出有用的量子计算应用,应用算法巧妙结合错误处理是必不可少的,期待更多这方面的工作。
Jared D. Weidman, Manas Sajjan, Camille Mikolas, Zachary J. Stewart, Johannes Pollanen, Sabre Kais, Angela K. Wilson
8.
尹志 (2025-09-30 22:39):
#paper Quantum computing and artificial intelligence: status and perspectives. doi: 10.48550/arXiv.2505.23860 比较新的一篇QAI的综述。比较细致的介绍了Quantum for AI及AI for Quantum,还有基础问题。最后介绍了一些目前这个领域所遇到的挑战。有两个特点值得一提,一个就是确实很新,目前基本的QAI的问题都有涉及;第二个就是这是一个全欧洲阵容的研究人员写的QAI综述,文章的开头就明确了自己的位置,这点还是很耐人寻味的。
arXiv, 2025-05-29T08:15:23Z. DOI: 10.48550/arXiv.2505.23860
Giovanni Acampora, Andris Ambainis, Natalia Ares, Leonardo Banchi, Pallavi Bhardwaj, Daniele Binosi, G. Andrew D. Briggs, Tommaso Calarco, Vedran Dunjko, Jens Eisert ... >>>
Giovanni Acampora, Andris Ambainis, Natalia Ares, Leonardo Banchi, Pallavi Bhardwaj, Daniele Binosi, G. Andrew D. Briggs, Tommaso Calarco, Vedran Dunjko, Jens Eisert, Olivier Ezratty, Paul Erker, Federico Fedele, Elies Gil-Fuster, Martin Gärttner, Mats Granath, Markus Heyl, Iordanis Kerenidis, Matthias Klusch, Anton Frisk Kockum, Richard Kueng, Mario Krenn, Jörg Lässig, Antonio Macaluso, Sabrina Maniscalco, Florian Marquardt, Kristel Michielsen, Gorka Muñoz-Gil, Daniel Müssig, Hendrik Poulsen Nautrup, Sophie A. Neubauer, Evert van Nieuwenburg, Roman Orus, Jörg Schmiedmayer, Markus Schmitt, Philipp Slusallek, Filippo Vicentini, Christof Weitenberg, Frank K. Wilhelm <<<
Abstract:
This white paper discusses and explores the various points of intersection
between quantum computing and artificial intelligence (AI). It describes how
quantum computing could support the development of innovative AI solutions. It
also examines use cases of classical AI that can empower research and
development in quantum technologies, with a focus on quantum computing and
quantum sensing. The purpose of this white paper is to provide a long-term
research agenda aimed at addressing foundational questions about how AI and
quantum computing interact and benefit one another.… >>>
This white paper discusses and explores the various points of intersection<br>between quantum computing and artificial intelligence (AI). It describes how<br>quantum computing could support the development of innovative AI solutions. It<br>also examines use cases of classical AI that can empower research and<br>development in quantum technologies, with a focus on quantum computing and<br>quantum sensing. The purpose of this white paper is to provide a long-term<br>research agenda aimed at addressing foundational questions about how AI and<br>quantum computing interact and benefit one another. It concludes with a set of<br>recommendations and challenges, including how to orchestrate the proposed<br>theoretical work, align quantum AI developments with quantum hardware roadmaps,<br>estimate both classical and quantum resources - especially with the goal of<br>mitigating and optimizing energy consumption - advance this emerging hybrid<br>software engineering discipline, and enhance European industrial<br>competitiveness while considering societal implications. <<<
9.
尹志 (2025-08-31 12:56):
#paper doi:10.48550/arXiv.2505.13683, ISCA, 2025, Genesis: A Compiler Framework for Hamiltonian Simulation on Hybrid CV-DV Quantum Computers. 作者引入了第一个基于连续离散混合量子计算系统的针对哈密顿量模拟的量子编译框架,非常有意思的工作。该框架分为哈密顿量初步分解和进一步的mapping和routing。也在几个常见的 物理模型上做了验证。量子编译作为量子计算机的一个重要环节,值得更多关注和技术的突破。
arXiv, 2025-05-19T19:32:06Z. DOI: 10.48550/arXiv.2505.13683
Zihan Chen, Jiakang Li, Minghao Guo, Henry Chen, Zirui Li, Joel Bierman, Yipeng Huang, Huiyang Zhou, Yuan Liu, Eddy Z. Zhang
Abstract:
This paper introduces Genesis, the first compiler designed to support
Hamiltonian Simulation on hybrid continuous-variable (CV) and discrete-variable
(DV) quantum computing systems. Genesis is a two-level compilation system. At
the first level, it decomposes an input Hamiltonian into basis gates using the
native instruction set of the target hybrid CV-DV quantum computer. At the
second level, it tackles the mapping and routing of qumodes/qubits to implement
long-range interactions for the gates decomposed from the first level. Rather
than a typical implementation that rel… >>>
This paper introduces Genesis, the first compiler designed to support<br>Hamiltonian Simulation on hybrid continuous-variable (CV) and discrete-variable<br>(DV) quantum computing systems. Genesis is a two-level compilation system. At<br>the first level, it decomposes an input Hamiltonian into basis gates using the<br>native instruction set of the target hybrid CV-DV quantum computer. At the<br>second level, it tackles the mapping and routing of qumodes/qubits to implement<br>long-range interactions for the gates decomposed from the first level. Rather<br>than a typical implementation that relies on SWAP primitives similar to<br>qubit-based (or DV-only) systems, we propose an integrated design of<br>connectivity-aware gate synthesis and beamsplitter SWAP insertion tailored for<br>hybrid CV-DV systems. We also introduce an OpenQASM-like domain-specific<br>language (DSL) named CVDV-QASM to represent Hamiltonian in terms of<br>Pauli-exponentials and basic gate sequences from the hybrid CV-DV gate set.<br>Genesis has successfully compiled several important Hamiltonians, including the<br>Bose-Hubbard model, $\mathbb{Z}_2-$Higgs model, Hubbard-Holstein model,<br>Heisenberg model and Electron-vibration coupling Hamiltonians, which are<br>critical in domains like quantum field theory, condensed matter physics, and<br>quantum chemistry. Our implementation is available at<br>Genesis-CVDV-Compiler(https://github.com/ruadapt/Genesis-CVDV-Compiler). <<<
10.
尹志 (2025-07-31 23:59):
#paper doi: 10.48550/arXiv.2507.06216 Unitary designs in nearly optimal depth. 文章设计了一种全新的量子电路,该电路可以接近理论最优深度高效构建unitray k-designs. 如果这个方案足够有效,那么对后续的量子算法的设计无疑非常有帮助。
arXiv, 2025-07-08T17:48:33Z. DOI: 10.48550/arXiv.2507.06216
Laura Cui, Thomas Schuster, Fernando Brandao, Hsin-Yuan Huang
Abstract:
We construct $\varepsilon$-approximate unitary $k$-designs on $n$ qubits in
circuit depth $O(\log k \log \log n k / \varepsilon)$. The depth is
exponentially improved over all known results in all three parameters $n$, $k$,
$\varepsilon$. We further show that each dependence is optimal up to
exponentially smaller factors. Our construction uses $\tilde{{O}}(nk)$ ancilla
qubits and ${O}(nk)$ bits of randomness, which are also optimal up to $\log(n
k)$ factors. An alternative construction achieves a smaller ancilla count
$\tilde{{O}}(n)$ with circuit depth ${O}(k \log \log n… >>>
We construct $\varepsilon$-approximate unitary $k$-designs on $n$ qubits in<br>circuit depth $O(\log k \log \log n k / \varepsilon)$. The depth is<br>exponentially improved over all known results in all three parameters $n$, $k$,<br>$\varepsilon$. We further show that each dependence is optimal up to<br>exponentially smaller factors. Our construction uses $\tilde{{O}}(nk)$ ancilla<br>qubits and ${O}(nk)$ bits of randomness, which are also optimal up to $\log(n<br>k)$ factors. An alternative construction achieves a smaller ancilla count<br>$\tilde{{O}}(n)$ with circuit depth ${O}(k \log \log nk/\varepsilon)$. To<br>achieve these efficient unitary designs, we introduce a highly-structured<br>random unitary ensemble that leverages long-range two-qubit gates and low-depth<br>implementations of random classical hash functions. We also develop a new<br>analytical framework for bounding errors in quantum experiments involving many<br>queries to random unitaries. As an illustration of this framework's<br>versatility, we provide a succinct alternative proof of the existence of<br>pseudorandom unitaries. <<<
11.
尹志 (2025-06-30 23:17):
#paper arXiv:2411.09131;Artificial Intelligence for Quantum Computing;2024;Yuri大佬带领的一篇综述,把AI用于量子计算的几个方面都做了分析和展望,虽然不是特别细致,但如果你希望量子计算能更快做出实际问题的优越性,显然不应该错过这篇综述。
arXiv, 2024-11-14T02:11:16Z. DOI: 10.48550/arXiv.2411.09131
Yuri Alexeev, Marwa H. Farag, Taylor L. Patti, Mark E. Wolf, Natalia Ares, Alán Aspuru-Guzik, Simon C. Benjamin, Zhenyu Cai, Zohim Chandani, Federico Fedele ... >>>
Yuri Alexeev, Marwa H. Farag, Taylor L. Patti, Mark E. Wolf, Natalia Ares, Alán Aspuru-Guzik, Simon C. Benjamin, Zhenyu Cai, Zohim Chandani, Federico Fedele, Nicholas Harrigan, Jin-Sung Kim, Elica Kyoseva, Justin G. Lietz, Tom Lubowe, Alexander McCaskey, Roger G. Melko, Kouhei Nakaji, Alberto Peruzzo, Sam Stanwyck, Norm M. Tubman, Hanrui Wang, Timothy Costa <<<
Abstract:
Artificial intelligence (AI) advancements over the past few years have had an
unprecedented and revolutionary impact across everyday application areas. Its
significance also extends to technical challenges within science and
engineering, including the nascent field of quantum computing (QC). The
counterintuitive nature and high-dimensional mathematics of QC make it a prime
candidate for AI's data-driven learning capabilities, and in fact, many of QC's
biggest scaling challenges may ultimately rest on developments in AI. However,
bringing leading techniques from AI to QC r… >>>
Artificial intelligence (AI) advancements over the past few years have had an<br>unprecedented and revolutionary impact across everyday application areas. Its<br>significance also extends to technical challenges within science and<br>engineering, including the nascent field of quantum computing (QC). The<br>counterintuitive nature and high-dimensional mathematics of QC make it a prime<br>candidate for AI's data-driven learning capabilities, and in fact, many of QC's<br>biggest scaling challenges may ultimately rest on developments in AI. However,<br>bringing leading techniques from AI to QC requires drawing on disparate<br>expertise from arguably two of the most advanced and esoteric areas of computer<br>science. Here we aim to encourage this cross-pollination by reviewing how<br>state-of-the-art AI techniques are already advancing challenges across the<br>hardware and software stack needed to develop useful QC - from device design to<br>applications. We then close by examining its future opportunities and obstacles<br>in this space. <<<
12.
尹志 (2025-05-31 21:23):
#paper https://doi.org/10.48550/arXiv.2012.07436 Informer: Beyond Efficient Transformer for Long Sequence Time-Series Forecasting。这是AAAI2021上的一篇关于长序列时序建模的经典工作。文章对传统Transformer进行了改进,提出了一类新的模型Informer,通过对self attention的改进和蒸馏,以及generative style decoder的构建,在时间复杂度、空间复杂度上都改善了传统Transformer存在的问题。该工作在多个数据集上取得了良好的性能。上述的几个思路在后续的时序建模中被频繁使用,非常具有启发性。
arXiv, 2020-12-14T11:43:09Z. DOI: 10.48550/arXiv.2012.07436
Haoyi Zhou, Shanghang Zhang, Jieqi Peng, Shuai Zhang, Jianxin Li, Hui Xiong, Wancai Zhang
Abstract:
Many real-world applications require the prediction of long sequence
time-series, such as electricity consumption planning. Long sequence
time-series forecasting (LSTF) demands a high prediction capacity of the model,
which is the ability to capture precise long-range dependency coupling between
output and input efficiently. Recent studies have shown the potential of
Transformer to increase the prediction capacity. However, there are several
severe issues with Transformer that prevent it from being directly applicable
to LSTF, including quadratic time complexity, high mem… >>>
Many real-world applications require the prediction of long sequence<br>time-series, such as electricity consumption planning. Long sequence<br>time-series forecasting (LSTF) demands a high prediction capacity of the model,<br>which is the ability to capture precise long-range dependency coupling between<br>output and input efficiently. Recent studies have shown the potential of<br>Transformer to increase the prediction capacity. However, there are several<br>severe issues with Transformer that prevent it from being directly applicable<br>to LSTF, including quadratic time complexity, high memory usage, and inherent<br>limitation of the encoder-decoder architecture. To address these issues, we<br>design an efficient transformer-based model for LSTF, named Informer, with<br>three distinctive characteristics: (i) a $ProbSparse$ self-attention mechanism,<br>which achieves $O(L \log L)$ in time complexity and memory usage, and has<br>comparable performance on sequences' dependency alignment. (ii) the<br>self-attention distilling highlights dominating attention by halving cascading<br>layer input, and efficiently handles extreme long input sequences. (iii) the<br>generative style decoder, while conceptually simple, predicts the long<br>time-series sequences at one forward operation rather than a step-by-step way,<br>which drastically improves the inference speed of long-sequence predictions.<br>Extensive experiments on four large-scale datasets demonstrate that Informer<br>significantly outperforms existing methods and provides a new solution to the<br>LSTF problem. <<<
13.
尹志 (2025-04-30 15:56):
#paper doi:10.48550/arXiv.2407.20516, Machine Unlearning in Generative AI: A Survey. 很有意思的方向,应该是翻译机器遗忘吧。随着模型越做越大,如何通过对模型的处理达到可控的添加与擦除特定信息,是未来一个重要的主题,不管是从隐私保护还是模型控制的层面上
arXiv, 2024-07-30T03:26:09Z. DOI: 10.48550/arXiv.2407.20516
Zheyuan Liu, Guangyao Dou, Zhaoxuan Tan, Yijun Tian, Meng Jiang
Abstract:
Generative AI technologies have been deployed in many places, such as
(multimodal) large language models and vision generative models. Their
remarkable performance should be attributed to massive training data and
emergent reasoning abilities. However, the models would memorize and generate
sensitive, biased, or dangerous information originated from the training data
especially those from web crawl. New machine unlearning (MU) techniques are
being developed to reduce or eliminate undesirable knowledge and its effects
from the models, because those that were designed for t… >>>
Generative AI technologies have been deployed in many places, such as<br>(multimodal) large language models and vision generative models. Their<br>remarkable performance should be attributed to massive training data and<br>emergent reasoning abilities. However, the models would memorize and generate<br>sensitive, biased, or dangerous information originated from the training data<br>especially those from web crawl. New machine unlearning (MU) techniques are<br>being developed to reduce or eliminate undesirable knowledge and its effects<br>from the models, because those that were designed for traditional<br>classification tasks could not be applied for Generative AI. We offer a<br>comprehensive survey on many things about MU in Generative AI, such as a new<br>problem formulation, evaluation methods, and a structured discussion on the<br>advantages and limitations of different kinds of MU techniques. It also<br>presents several critical challenges and promising directions in MU research. A<br>curated list of readings can be found:<br>https://github.com/franciscoliu/GenAI-MU-Reading. <<<
14.
尹志 (2025-03-31 15:06):
#paper:doi:doi.org/10.48550/arXiv.2502.11974, Image Inversion: A Survey from GANs to Diffusion and Beyond(2025). 综述了image inversion常见的算法模型,很新,主要介绍了GAN和diffusion模型,也提了DiT和Rectified Flow框架。image inversion的核心问题涉及latent space, 对其它生成式AI的问题都非常重要。
arXiv, 2025-02-17T16:20:48Z. DOI: 10.48550/arXiv.2502.11974
Yinan Chen, Jiangning Zhang, Yali Bi, Xiaobin Hu, Teng Hu, Zhucun Xue, Ran Yi, Yong Liu, Ying Tai
Abstract:
Image inversion is a fundamental task in generative models, aiming to map
images back to their latent representations to enable downstream applications
such as editing, restoration, and style transfer. This paper provides a
comprehensive review of the latest advancements in image inversion techniques,
focusing on two main paradigms: Generative Adversarial Network (GAN) inversion
and diffusion model inversion. We categorize these techniques based on their
optimization methods. For GAN inversion, we systematically classify existing
methods into encoder-based approaches, lat… >>>
Image inversion is a fundamental task in generative models, aiming to map<br>images back to their latent representations to enable downstream applications<br>such as editing, restoration, and style transfer. This paper provides a<br>comprehensive review of the latest advancements in image inversion techniques,<br>focusing on two main paradigms: Generative Adversarial Network (GAN) inversion<br>and diffusion model inversion. We categorize these techniques based on their<br>optimization methods. For GAN inversion, we systematically classify existing<br>methods into encoder-based approaches, latent optimization approaches, and<br>hybrid approaches, analyzing their theoretical foundations, technical<br>innovations, and practical trade-offs. For diffusion model inversion, we<br>explore training-free strategies, fine-tuning methods, and the design of<br>additional trainable modules, highlighting their unique advantages and<br>limitations. Additionally, we discuss several popular downstream applications<br>and emerging applications beyond image tasks, identifying current challenges<br>and future research directions. By synthesizing the latest developments, this<br>paper aims to provide researchers and practitioners with a valuable reference<br>resource, promoting further advancements in the field of image inversion. We<br>keep track of the latest works at https://github.com/RyanChenYN/ImageInversion <<<
15.
尹志 (2025-02-28 15:55):
#paper doi:10.48550/arXiv.2205.15463 Few-Shot Diffusion Models. 文章提出了一种扩散模型及set-based ViT的方式实现few shot生成的技术。实验表明,该模型仅需5个样本就可以完成新类别的生成。
arXiv, 2022-05-30T23:20:33Z. DOI: 10.48550/arXiv.2205.15463
Giorgio Giannone, Didrik Nielsen, Ole Winther
Abstract:
Denoising diffusion probabilistic models (DDPM) are powerful hierarchical
latent variable models with remarkable sample generation quality and training
stability. These properties can be attributed to parameter sharing in the
generative hierarchy, as well as a parameter-free diffusion-based inference
procedure. In this paper, we present Few-Shot Diffusion Models (FSDM), a
framework for few-shot generation leveraging conditional DDPMs. FSDMs are
trained to adapt the generative process conditioned on a small set of images
from a given class by aggregating image patch inform… >>>
Denoising diffusion probabilistic models (DDPM) are powerful hierarchical<br>latent variable models with remarkable sample generation quality and training<br>stability. These properties can be attributed to parameter sharing in the<br>generative hierarchy, as well as a parameter-free diffusion-based inference<br>procedure. In this paper, we present Few-Shot Diffusion Models (FSDM), a<br>framework for few-shot generation leveraging conditional DDPMs. FSDMs are<br>trained to adapt the generative process conditioned on a small set of images<br>from a given class by aggregating image patch information using a set-based<br>Vision Transformer (ViT). At test time, the model is able to generate samples<br>from previously unseen classes conditioned on as few as 5 samples from that<br>class. We empirically show that FSDM can perform few-shot generation and<br>transfer to new datasets. We benchmark variants of our method on complex vision<br>datasets for few-shot learning and compare to unconditional and conditional<br>DDPM baselines. Additionally, we show how conditioning the model on patch-based<br>input set information improves training convergence. <<<
16.
尹志 (2025-01-31 17:05):
#paper https://doi.org/10.48550/arXiv.2403.07183 Monitoring AI-Modified Content at Scale: A Case Study on the Impact of ChatGPT on AI Conference Peer Reviews 一篇讨论大语言模型使用情况的文章, 特别举了在AI顶会评审中使用的具体例子。(包括ICLR 2024、NeurIPS 2023、CoRL 2023和EMNLP 2023。)研究发现,这些论文review中,有6.5%至16.9%可能被LLM大幅修改,而且这些review有很多有趣的特点,比如confidence比较低,接近ddl才提交,而且不太愿意回应作者反驳等。更多有趣的现象可参考原文。文章中贴了最常见的AI喜欢使用的形容词,比如“commendable”, “meticulous”, and “intricate”等,确实很像AI搞的,哈哈哈。 看来以后审稿人要对作者更加负责才行噢。
arXiv, 2024-03-11T21:51:39Z. DOI: 10.48550/arXiv.2403.07183
Weixin Liang, Zachary Izzo, Yaohui Zhang, Haley Lepp, Hancheng Cao, Xuandong Zhao, Lingjiao Chen, Haotian Ye, Sheng Liu, Zhi Huang ... >>>
Weixin Liang, Zachary Izzo, Yaohui Zhang, Haley Lepp, Hancheng Cao, Xuandong Zhao, Lingjiao Chen, Haotian Ye, Sheng Liu, Zhi Huang, Daniel A. McFarland, James Y. Zou <<<
Abstract:
We present an approach for estimating the fraction of text in a large corpus
which is likely to be substantially modified or produced by a large language
model (LLM). Our maximum likelihood model leverages expert-written and
AI-generated reference texts to accurately and efficiently examine real-world
LLM-use at the corpus level. We apply this approach to a case study of
scientific peer review in AI conferences that took place after the release of
ChatGPT: ICLR 2024, NeurIPS 2023, CoRL 2023 and EMNLP 2023. Our results suggest
that between 6.5% and 16.9% of text submitted … >>>
We present an approach for estimating the fraction of text in a large corpus<br>which is likely to be substantially modified or produced by a large language<br>model (LLM). Our maximum likelihood model leverages expert-written and<br>AI-generated reference texts to accurately and efficiently examine real-world<br>LLM-use at the corpus level. We apply this approach to a case study of<br>scientific peer review in AI conferences that took place after the release of<br>ChatGPT: ICLR 2024, NeurIPS 2023, CoRL 2023 and EMNLP 2023. Our results suggest<br>that between 6.5% and 16.9% of text submitted as peer reviews to these<br>conferences could have been substantially modified by LLMs, i.e. beyond<br>spell-checking or minor writing updates. The circumstances in which generated<br>text occurs offer insight into user behavior: the estimated fraction of<br>LLM-generated text is higher in reviews which report lower confidence, were<br>submitted close to the deadline, and from reviewers who are less likely to<br>respond to author rebuttals. We also observe corpus-level trends in generated<br>text which may be too subtle to detect at the individual level, and discuss the<br>implications of such trends on peer review. We call for future<br>interdisciplinary work to examine how LLM use is changing our information and<br>knowledge practices. <<<
17.
尹志 (2024-12-29 21:50):
#paper doi: 10.1021/acs.jcim.4c01107 Journal of Chemical Information and Modeling, 2024, Diffusion Models in De Novo Drug Design。又是一篇diffusion model用于药物设计的综述,很新很全面。这个领域的发展特别快,各种方法层出不穷,有了诺奖这次的加持,应该后面几年会有更多务实的成果。
18.
尹志 (2024-11-30 22:05):
#paper https://doi.org/10.48550/arXiv.1701.08223 2017, The Python-based Simulations of Chemistry Framework (PySCF)。非常重要的量子化学工具PySCF的介绍。2014年启动的项目,从一开始的仅仅有几个函数功能,到现在对各种量化问题的计算的良好支持,其易用性及可扩展性得到了社群的认可。这个特性其实在软件于2015年发布的时候就设定好了。因此,几乎所有功能代码都由python实现,只有遇到特别的time-ciritical的代码部分才去用c实现。当然,这个特性使得目前大量量化计算的库都依赖于pyscf,俨然成为开源领域的gaussion的有力竞争者。
arXiv, 2017-01-27T23:57:43Z. DOI: 10.48550/arXiv.1701.08223
Qiming Sun, Timothy C. Berkelbach, Nick S. Blunt, George H. Booth, Sheng Guo, Zhendong Li, Junzi Liu, James McClain, Elvira R. Sayfutyarova, Sandeep Sharma ... >>>
Qiming Sun, Timothy C. Berkelbach, Nick S. Blunt, George H. Booth, Sheng Guo, Zhendong Li, Junzi Liu, James McClain, Elvira R. Sayfutyarova, Sandeep Sharma, Sebastian Wouters, Garnet Kin-Lic Chan <<<
Abstract:
PySCF is a general-purpose electronic structure platform designed from the
ground up to emphasize code simplicity, both to aid new method development, as
well as for flexibility in computational workflow. The package provides a wide
range of tools to support simulations of finite size systems, extended systems
with periodic boundary conditions, low dimensional periodic systems, and custom
Hamiltonians, using mean-field and post-mean-field methods with standard
Gaussian basis functions. To ensure easy of extensibility, PySCF uses the
Python language to implement almost all… >>>
PySCF is a general-purpose electronic structure platform designed from the<br>ground up to emphasize code simplicity, both to aid new method development, as<br>well as for flexibility in computational workflow. The package provides a wide<br>range of tools to support simulations of finite size systems, extended systems<br>with periodic boundary conditions, low dimensional periodic systems, and custom<br>Hamiltonians, using mean-field and post-mean-field methods with standard<br>Gaussian basis functions. To ensure easy of extensibility, PySCF uses the<br>Python language to implement almost all its features, while computationally<br>critical paths are implemented with heavily optimized C routines. Using this<br>combined Python/C implementation, the package is as efficient as the best<br>existing C or Fortran based quantum chemistry programs. In this paper we<br>document the capabilities and design philosophy of the current version of the<br>PySCF package. <<<
19.
尹志 (2024-10-31 13:55):
#paper doi.org/10.1038/sdata.2014.22 Quantum chemistry structures and properties of 134 kilo molecules, Scientific Data 1, 140022, 2014. 这是著名的数据集QM9的原始论文,最近在做相关计算工作, 又好好读了一下。非常重要的工作,给后续各种量化计算提供了特别方便的benchmark。该工作使用DFT方法(B3LYP/6-31G(2df,p))计算了134k种小分子的各种量化性质,比如能量、偶极矩、极化率等。
Raghunathan Ramakrishnan, Pavlo O. Dral, Matthias Rupp, O. Anatole von Lilienfeld
20.
尹志 (2024-09-30 23:02):
#paper https://doi.org/10.48550/arXiv.2405.20328 mRNA secondary structure prediction using utility-scale quantum computers。 这是今年IBM和Moderna合作的一篇工作。作者用CVaR-based VQE算法对mRNA的二级结构做了预测。RNA由于其单链多变的特性,非常难以预测。当然也正是这个原因,在计算上很容易被归类到组合优化问题的范畴。因此利用量子计算机去设计特定算法来加速解决,并给出最优结构显得顺理成章。文章使用了IBM的量子处理器Eagle和Heron, 得出的结果和经典算法CPLEX保持一致。当然,考虑到使用了NISQ的方式,如何保证机器的校准及错误抑制文章并没有交代的很细致,默认Eagle和Heron已经做到了吧。当然,这也给VQC算法(包括VQE、QAOA)解决组合优化问题做了一个很好的示范,充分证明了变分算法的灵活性。
arXiv, 2024-05-30T17:58:17Z. DOI: 10.48550/arXiv.2405.20328
Dimitris Alevras, Mihir Metkar, Takahiro Yamamoto, Vaibhaw Kumar, Triet Friedhoff, Jae-Eun Park, Mitsuharu Takeori, Mariana LaDue, Wade Davis, Alexey Galda
Abstract:
Recent advancements in quantum computing have opened new avenues for tackling
long-standing complex combinatorial optimization problems that are intractable
for classical computers. Predicting secondary structure of mRNA is one such
notoriously difficult problem that can benefit from the ever-increasing
maturity of quantum computing technology. Accurate prediction of mRNA secondary
structure is critical in designing RNA-based therapeutics as it dictates
various steps of an mRNA life cycle, including transcription, translation, and
decay. The current generation of quantum … >>>
Recent advancements in quantum computing have opened new avenues for tackling<br>long-standing complex combinatorial optimization problems that are intractable<br>for classical computers. Predicting secondary structure of mRNA is one such<br>notoriously difficult problem that can benefit from the ever-increasing<br>maturity of quantum computing technology. Accurate prediction of mRNA secondary<br>structure is critical in designing RNA-based therapeutics as it dictates<br>various steps of an mRNA life cycle, including transcription, translation, and<br>decay. The current generation of quantum computers have reached utility-scale,<br>allowing us to explore relatively large problem sizes. In this paper, we<br>examine the feasibility of solving mRNA secondary structures on a quantum<br>computer with sequence length up to 60 nucleotides representing problems in the<br>qubit range of 10 to 80. We use Conditional Value at Risk (CVaR)-based VQE<br>algorithm to solve the optimization problems, originating from the mRNA<br>structure prediction problem, on the IBM Eagle and Heron quantum processors. To<br>our encouragement, even with ``minimal'' error mitigation and fixed-depth<br>circuits, our hardware runs yield accurate predictions of minimum free energy<br>(MFE) structures that match the results of the classical solver CPLEX. Our<br>results provide sufficient evidence for the viability of solving mRNA structure<br>prediction problems on a quantum computer and motivate continued research in<br>this direction. <<<
回到顶部