Eden’s Reading List Header

DeepSeek-V3.2: Pushing the Frontier of Open Large Language Models

Read More
Research & Engineering: Aixin Liu, Aoxue Mei, Bangcai Lin, Bing Xue, Bingxuan Wang, Bingzheng Xu, Bochao Wu, Bowei Zhang, Chaofan Lin, Chen Dong, Chengda Lu, Chenggang Zhao, Chengqi Deng, Chenhao Xu, Chong Ruan*, Damai Dai, Daya Guo, Dejian Yang, Deli Chen, Erhang Li, Fangqi Zhou*, Fangyun Lin, Fucong Dai, Guangbo Hao, Guanting Chen, Guowei Li, H. Zhang, Hanwei Xu, Hao Li, Haofen Liang, Haoran Wei, Haowei Zhang, Haowen Luo, Haozhe Ji, Honghui Ding, Hongxuan Tang, Huanqi Cao, Huazuo Gao, Hui Qu, Hui Zeng, Jialiang Huang, Jiashi Li, Jiaxin Xu, Jiewen Hu, Jingchang Chen, Jingting Xiang, Jingyang Yuan, Jingyuan Cheng, Jinhua Zhu, Jun Ran*, Junguang Jiang, Junjie Qiu, Junlong Li*, Junxiao Song, Kai Dong, Kaige Gao, Kang Guan, Kexin Huang*, Kexing Zhou, Kezhao Huang, Kuai Yu, Lean Wang, Lecong Zhang, Lei Wang, Liang Zhao, Liangsheng Yin*, Lihua Guo, Lingxiao Luo, Linwang Ma, Litong Wang, Liyue Zhang, M.S. Di, M.Y Xu, Mingchuan Zhang, Minghua Zhang, Minghui Tang, Mingxu Zhou, Panpan Huang, Peixin Cong, Peiyi Wang, Qiancheng Wang, Qihao Zhu, Qingyang Li, Qinyu Chen, Qiushi Du, Ruiling Xu, Ruiqi Ge, Ruisong Zhang, Ruizhe Pan, Runji Wang, Runqiu Yin, Runxin Xu, Ruomeng Shen, Ruoyu Zhang, S.H. Liu, Shanghao Lu, Shangyan Zhou, Shanhuang Chen, Shaofei Cai, Shaoyuan Chen, Shengding Hu, Shengyu Liu, Shiqiang Hu, Shirong Ma, Shiyu Wang, Shuiping Yu, Shunfeng Zhou, Shuting Pan, Songyang Zhou, Tao Ni, Tao Yun, Tian Pei, Tian Ye, Tianyuan Yue, Wangding Zeng, Wen Liu, Wenfeng Liang, Wenjie Pang, Wenjing Luo, Wenjun Gao, Wentao Zhang, Xi Gao, Xiangwen Wang, Xiao Bi, Xiaodong Liu, Xiaohan Wang, Xiaokang Chen, Xiaokang Zhang, Xiaotao Nie, Xin Cheng, Xin Liu, Xin Xie, Xingchao Liu, Xingkai Yu, Xingyou Li, Xinyu Yang, Xinyuan Li*, Xu Chen, Xuecheng Su, Xuehai Pan, Xuheng Lin, Xuwei Fu, Y.Q. Wang, Yang Zhang, Yanhong Xu, Yanru Ma, Yao Li, Yao Li, Yao Zhao, Yaofeng Sun, Yaohui Wang, Yi Qian, Yi Yu, Yichao Zhang, Yifan Ding, Yifan Shi, Yiliang Xiong, Ying He, Ying Zhou, Yinmin Zhong, Yishi Piao, Yisong Wang, Yixiao Chen, Yixuan Tan, Yixuan Wei, Yiyang Ma, Yiyuan Liu, Yonglun Yang, Yongqiang Guo, Yongtong Wu, Yu Wu, Yuan Cheng, Yuan Ou, Yuanfan Xu, Yuduan Wang, Yue Gong*, Yuhan Wu, Yuheng Zou, Yukun Li, Yunfan Xiong, Yuxiang Luo, Yuxiang You, Yuxuan Liu, Yuyang Zhou, Z.F. Wu, Z.Z. Ren, Zehua Zhao, Zehui Ren, Zhangli Sha, Zhe Fu, Zhean Xu, Zhenda Xie, Zhengyan Zhang, Zhewen Hao, Zhibin Gou, Zhicheng Ma, Zhigang Yan, Zhihong Shao, Zhixian Huang, Zhiyu Wu, Zhuoshu Li, Zhuping Zhang, Zian Xu, Zihao Wang, Zihui Gu, Zijia Zhu, Zilin Li, Zipeng Zhang, Ziwei Xie, Ziyi Gao, Zizheng Pan, Zongqing Yao Data Annotation: Bei Feng, Hui Li, J.L. Cai, Jiaqi Ni, Lei Xu, Meng Li, Ning Tian, R.J. Chen, R.L. Jin, S.S. Li, Shuang Zhou, Tianyu Sun, X.Q. Li, Xiangyue Jin, Xiaojin Shen, Xiaosha Chen, Xinnan Song, Xinyi Zhou, Y.X. Zhu, Yanping Huang, Yaohui Li, Yi Zheng, Yuchen Zhu, Yunxian Ma, Zhen Huang, Zhipeng Xu, Zhongyu Zhang Business & Compliance: Dongjie Ji, Jian Liang, Jianzhong Guo, Jin Chen, Leyi Xia, Miaojun Wang, Mingming Li, Peng Zhang, Ruyi Chen, Shangmian Sun, Shaoqing Wu, Shengfeng Ye, T.Wang, W.L. Xiao, Wei An, Xianzu Wang, Xiaowen Sun, Xiaoxiang Wang, Ying Tang, Yukun Zha, Zekai Zhang, Zhe Ju, Zhen Zhang, Zihua Qu
December 8, 2025
DeepSeek-V3.2 is an advanced open-source large language model (LLM) that introduces a highly efficient sparse attention mechanism (DSA) and scalable reinforcement learning (RL) techniques, achieving reasoning and agent-task performance comparable to GPT-5 and close to the closed-source leader Gemini-3.0-Pro, including gold-medal level results in top coding and math competitions like IOI 2025 and IMO 2025; it significantly narrows the gap between open and proprietary LLMs on complex reasoning, coding (relevant to Aleph portfolio company Grain Finance’s interest in AI/code workflows), and tool-use benchmarks (potentially impacting Aleph portfolio companies such as Superlegal and Ply that involve AI agents/tool integration), while emphasizing cost-efficient deployment, improved long-context handling, and novel synthetic environment generation for RL training that enhances agent generalization in interactive real-world tasks.

Latent Collaboration in Multi-Agent Systems

Read More
Jiaru Zou Princeton University University of Illinois Urbana-Champaign Co-Leadership Core Contributors, Xiyuan Yang University of Illinois Urbana-Champaign Co-Leadership Core Contributors, Ruizhong Qiu University of Illinois Urbana-Champaign Core Contributors, Gaotang Li University of Illinois Urbana-Champaign Core Contributors, Katherine Tieu University of Illinois Urbana-Champaign Core Contributors, Pan Lu Stanford University Core Contributors, Ke Shen Hanghang Tong University of Illinois Urbana-Champaign, Yejin Choi Stanford University, Jingrui He University of Illinois Urbana-Champaign, James Zou Stanford University, Mengdi Wang Princeton University, Ling Yang Princeton University
December 3, 2025
A new multi-agent system framework LatentMAS enables large language models (LLMs) to collaborate entirely in continuous latent space rather than text, achieving up to 14.6% higher accuracy, 4x faster inference, and ~70-84% reduction in token usage across diverse benchmarks including math, science, commonsense reasoning, and code generation; compared to conventional text-based multi-agent collaboration methods, LatentMAS provides lossless, more expressive, and computationally efficient cross-agent communication via shared latent working memory (KV caches) without additional training, offering a promising new paradigm for agentic AI systems relevant to portfolio companies like Anodot and Sequence that rely on advanced AI reasoning, or Ply and Simply that could benefit from more efficient multi-agent LLM collaboration frameworks.

An updated evolutionary classification of CRISPR–Cas systems including rare variants | Nature Microbiology

Read More
Kira S. Makarova, Sergey A. Shmakov, Yuri I. Wolf, Pascal Mutz, Han Altae-Tran, Chase L. Beisel, Stan J. J. Brouns, Emmanuelle Charpentier, David Cheng, Jennifer Doudna, Daniel H. Haft, Philippe Horvath, Sylvain Moineau, Francisco J. M. Mojica, Patrick Pausch, Rafael Pinilla-Redondo, Shiraz A. Shah, Virginijus Siksnys, Michael P. Terns, Jesse Tordoff, Česlovas Venclovas, Malcolm F. White, Alexander F. Yakunin, Feng Zhang, Eugene V. Koonin
November 17, 2025
A major update in the evolutionary classification of CRISPR–Cas systems expands the framework to 2 classes, 7 types, and 46 subtypes (from 6 types and 33 subtypes five years ago), incorporating newly identified rare variants—particularly multiple new class 1 subtypes (III-G, III-H, III-I) and a distinct type VII with Cas14 nuclease—reflecting a broad diversity mainly involved in prokaryotic adaptive immunity and defense against viruses and mobile genetic elements (MGEs). The analysis highlights extensive modular evolution, signaling pathway diversity (cOA and SAM–AMP second messengers in type III), and reveals frequent recruitment of CRISPR–Cas by transposons for RNA-guided transposition, plus repeated exaptation for non-defense functions. Distribution studies show CRISPR–Cas systems are highly prevalent in archaea (especially thermophiles) and bacteria, with class 1 systems generally dominating over class 2. Newly discovered rare variants contribute only about 0.3% of systems in sequenced genomes but emphasize a long tail of diversity needing further mining aided by AI and metagenomic data. For Aleph portfolio companies operating in AI, SaaS for data-driven products, and cybersecurity domains such as Anodot, Panorays, Ply, Sequence, and Superlegal, these insights indicate a rapidly evolving landscape of CRISPR-related molecular tools and defense mechanisms—highlighting potential areas for advanced genomic data analytics and biosecurity applications, and possibly informing innovative product development leveraging CRISPR technologies or their evolutionary dynamics.

AI Interview Series #1: Explain Some LLM Text Generation Strategies Used in LLMs - MarkTechPost

Read More
Arham Islam
November 17, 2025
The article explains key large language model (LLM) text generation strategies—Greedy Search, Beam Search, Nucleus (Top-p) Sampling, and Temperature Sampling—highlighting their differing balance of focus, creativity, and coherence, relevant for Aleph portfolio companies using or developing LLM-based products like LawGeex (legal AI) and Superlegal (legal workflows), where precise vs. varied text generation matters. It also presents a detailed PyTorch implementation of a continual learning neural memory agent combining differentiable memory, experience replay, and meta-learning for adaptive AI systems, which could inform Aleph investments in AI infrastructure like NextSilicon or Grain Finance that may require stable continual learning models. Additionally, Kosmos—a new autonomous AI scientific discovery system running extensive research cycles using a structured long-term world model and multi-agent design—is described in depth; it reads thousands of papers and runs tens of thousands of lines of code to generate credible scientific reports with ~79% accuracy, reproducing known results and proposing novel mechanisms across domains including metabolomics, materials science, neuroscience, and genetics. This breakthrough in AI-augmented research demonstrates advanced LLM agent orchestration and reasoning capabilities with reproducibility and auditability, relevant to Aleph portfolio sectors like Windward (maritime AI leveraging complex data) and Fabric (supply chain automation), where integrating structured memory and multi-agent workflows can boost decision-making. Finally, the article surveys common memory architectures for multi-agent LLM systems, highlighting vector memory systems (e.g., vector retrieval augmented generation) as the dominant pattern for fast, scalable but limited memory suited for local or embedded queries, while noting their weaknesses on temporal, cross-session, or multi-hop relational tasks—insights critical for building robust AI agents like those in Sightful (sales intelligence) or Unit (embedded banking), where memory design impacts system reliability and user experience.

ReasoningBank: Scaling Agent Self-Evolving with Reasoning Memory

Read More
Siru Ouyang, Jun Yan, I-Hung Hsu, Yanfei Chen, Ke Jiang, Zifeng Wang, Rujun Han, Long T. Le, Samira Daruki, Xiangru Tang, Vishy Tirumalashetty, George Lee, Mahsan Rofouei, Hangfei Lin, Jiawei Han, Chen-Yu Lee, Tomas Pfister
October 26, 2025
The paper introduces ReasoningBank, a novel memory framework for large language model (LLM) agents that distills transferable reasoning strategies from both successful and failed experiences to enable continuous self-improvement in persistent real-world tasks like web browsing and software engineering; combined with a new memory-aware test-time scaling method (MaTTS), it significantly outperforms existing memory mechanisms and scaling approaches in benchmarks, improving effectiveness and efficiency by leveraging rich, structured memories rather than just raw trajectories or successful workflows. This advancement is relevant for portfolio companies like Ply (ply.io) and Sequence (getsequence.io), which build agent-driven automation and workflow tools that could benefit from enhanced agent memory and scaling capabilities; it also impacts competitors and partners in the LLM agent ecosystem, such as those developing autonomous web or code agents (e.g., WebArena, BrowserGym environments). The synergy between memory and test-time compute scaling demonstrated by ReasoningBank and MaTTS illustrates a practical path toward adaptive, lifelong-learning AI agents with emergent reasoning strategies.

The astrocytic ensemble acts as a multiday trace to stabilize memory | Nature

Read More
Ken-ichi Dewa, Kodai Kaseda, Aoi Kuwahara, Hideaki Kubotera, Ayato Yamasaki, Natsumi Awata, Atsuko Komori, Mika A. Holtz, Atsushi Kasai, Henrik Skibbe, Norio Takata, Tatsushi Yokoyama, Makoto Tsuda, Genri Numata, Shun Nakamura, Eiki Takimoto, Masayuki Sakamoto, Minako Ito, Takahiro Masuda & Jun Nagai
October 26, 2025
This comprehensive study published in Nature reveals that astrocytes form behaviorally relevant ensembles (BAEs) acting as multiday molecular traces critical for stabilizing emotionally salient, repeated fear memories, especially in the amygdala—a key region implicated in fear memory—by integrating local engram neuronal activity with long-range noradrenaline (NA) signaling. Using novel brain-wide astrocyte-specific Fos tagging combined with imaging and transcriptomics, the authors demonstrate that initial fear conditioning (FC) primes astrocytes by upregulating α1- and β1-adrenoreceptors over a day, enhancing their responsiveness to NA during fear recall (FR), which activates Fos and induces neuromodulatory IGFBP2 expression. Perturbation of astrocytic NA signaling or IGFBP2 impairs memory restabilization/reconsolidation, while astrocyte β1-adrenergic receptor overexpression amplifies astrocyte ensembles and leads to memory over-stabilization and generalization, as evidenced by increased reactivation of neuronal engram cells during recall. This astrocyte ensemble's slower timescale of activity contrasts with the rapid neuronal engrams and suggests astrocytes provide eligibility or stabilization traces over days to maintain memory precision, highlighting their integral role in memory circuits beyond neurons. These findings have implications for disorders with dysregulated noradrenergic signaling, such as PTSD, and open avenues for targeting astrocyte–neuron interactions in cognitive and neuropsychiatric therapeutics. Although none of Aleph’s portfolio companies are directly mentioned, the study’s insights into neural circuit stability, neuromodulation, and state-dependent molecular ensembles may be relevant in the broader neurotechnology and AI-driven cognitive analytics domains related to companies like Grain Finance (cognitive modeling of economic memory) or Sightful (interpreting nuanced human behavior), and intersect with AI and data analytics innovations in brain-inspired computing.