当前共找到 1086 篇文献分享,本页显示第 661 - 680 篇。
661.
李翛然 (2022-10-31 09:48):
#paper TET1 is a beige adipocyte-selective epigenetic suppressor of thermogenesis doi: https://doi.org/10.1038/s41467-020-18054-y 关于TET1 ,文献报道Tumor suppressor应该是没有问题,做为重组蛋白治疗肿瘤,我接下来要调研一下临床上哪类肿瘤病人是否有TET1缺失的现象,由此来判断肿瘤是否在TET1不缺失的情况下不好生长,确定其临床价值,还有一个要考虑的是这2篇文章介绍的TET1压抑脂肪细胞热能量代谢,维他命C作用在TET1压制somatic cell reprogramming,这2个现象是否可能导致严重的副作用,限制TET1的剂量
IF:14.700Q1 Nature communications, 2020-08-27. DOI: 10.1038/s41467-020-18054-y PMID: 32855402
Abstract:
It has been suggested that beige fat thermogenesis is tightly controlled by epigenetic regulators that sense environmental cues such as temperature. Here, we report that subcutaneous adipose expression of the … >>>
It has been suggested that beige fat thermogenesis is tightly controlled by epigenetic regulators that sense environmental cues such as temperature. Here, we report that subcutaneous adipose expression of the DNA demethylase TET1 is suppressed by cold and other stimulators of beige adipocyte thermogenesis. TET1 acts as an autonomous repressor of key thermogenic genes, including Ucp1 and Ppargc1a, in beige adipocytes. Adipose-selective Tet1 knockout mice generated by using Fabp4-Cre improves cold tolerance and increases energy expenditure and protects against diet-induced obesity and insulin resistance. Moreover, the suppressive role of TET1 in the thermogenic gene regulation of beige adipocytes is largely DNA demethylase-independent. Rather, TET1 coordinates with HDAC1 to mediate the epigenetic changes to suppress thermogenic gene transcription. Taken together, TET1 is a potent beige-selective epigenetic breaker of the thermogenic gene program. Our findings may lead to a therapeutic strategy to increase energy expenditure in obesity and related metabolic disorders. <<<
翻译
662.
前进 (2022-10-30 21:26):
#paper Shi J, He Y, Kong Y, et al. XMorpher: Full Transformer for Deformable Medical Image Registration via Cross Attention[C]//International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer, Cham, 2022: 217-226.现有的深度网络专注于单个图像的特征提取,并且在对成对图像执行的配准任务方面受到限制。因此,本文提出了一种新的骨干网络XMorpher,有效地对变形配准中成对特征进行表示。1) 它提出了一种新的Transformer架构,包括双并行特征提取网络,该网络通过Cross Attention来改变信息,从而发现多级语义对应关系,同时逐步提取各自的特征,以实现最终的有效配准。2) 它提出了Cross Attention Transformer(CAT)块,以建立图像之间的注意力机制,该机制能够自动找到对应关系,并促使特征在网络中有效融合。3) 它限制了不同大小的基本窗口和搜索窗口之间的计算,从而集中于可变形配准的局部变换,同时提高了计算效率。XMorpher使Voxelmorph在DSC上提高了2.8%,证明了其在变形配准中对配对图像的特征的有效表示。
Abstract:
An effective backbone network is important to deep learning-based Deformable Medical Image Registration (DMIR), because it extracts and matches the features between two images to discover the mutual correspondence for … >>>
An effective backbone network is important to deep learning-based Deformable Medical Image Registration (DMIR), because it extracts and matches the features between two images to discover the mutual correspondence for fine registration. However, the existing deep networks focus on single image situation and are limited in registration task which is performed on paired images. Therefore, we advance a novel backbone network, XMorpher, for the effective corresponding feature representation in DMIR. 1) It proposes a novel full transformer architecture including dual parallel feature extraction networks which exchange information through cross attention, thus discovering multi-level semantic correspondence while extracting respective features gradually for final effective registration. 2) It advances the Cross Attention Transformer (CAT) blocks to establish the attention mechanism between images which is able to find the correspondence automatically and prompts the features to fuse efficiently in the network. 3) It constrains the attention computation between base windows and searching windows with different sizes, and thus focuses on the local transformation of deformable registration and enhances the computing efficiency at the same time. Without any bells and whistles, our XMorpher gives Voxelmorph 2.8% improvement on DSC, demonstrating its effective representation of the features from the paired images in DMIR. We believe that our XMorpher has great application potential in more paired medical images. Our XMorpher is open on https://github.com/Solemoon/XMorpher <<<
翻译
663.
洪媛媛 (2022-10-30 12:16):
#paper https://doi.org/10.1038/s41467-022-32995-6 nature communications 2022. Cost-effective methylome sequencing of cell-free DNA for accurately detecting and locating cancer. 这篇文章介绍了一种富集cfDNA CpG区域的NGS建库方法(cfMethyl-Seq),cfMethyl-Seq比全基因组甲基化测序更节省数据量,而且比传统的RRBS方法更适合用于cfDNA CpG区域的富集。该研究首先通过RRBS测序的癌症、癌旁组织样本,以及cfMethyl-Seq测序的健康人血浆样本,筛选出癌症早筛和组织溯源(TOO)marker,然后将cfMethyl-Seq测序的217癌症和131健康人血浆样本,分成训练集和测试集,在训练集建模,在测试集验证性能。
IF:14.700Q1 Nature communications, 2022-09-29. DOI: 10.1038/s41467-022-32995-6 PMID: 36175411
Abstract:
Early cancer detection by cell-free DNA faces multiple challenges: low fraction of tumor cell-free DNA, molecular heterogeneity of cancer, and sample sizes that are not sufficient to reflect diverse patient … >>>
Early cancer detection by cell-free DNA faces multiple challenges: low fraction of tumor cell-free DNA, molecular heterogeneity of cancer, and sample sizes that are not sufficient to reflect diverse patient populations. Here, we develop a cancer detection approach to address these challenges. It consists of an assay, cfMethyl-Seq, for cost-effective sequencing of the cell-free DNA methylome (with > 12-fold enrichment over whole genome bisulfite sequencing in CpG islands), and a computational method to extract methylation information and diagnose patients. Applying our approach to 408 colon, liver, lung, and stomach cancer patients and controls, at 97.9% specificity we achieve 80.7% and 74.5% sensitivity in detecting all-stage and early-stage cancer, and 89.1% and 85.0% accuracy for locating tissue-of-origin of all-stage and early-stage cancer, respectively. Our approach cost-effectively retains methylome profiles of cancer abnormalities, allowing us to learn new features and expand to other cancer types as training cohorts grow. <<<
翻译
664.
负负 (2022-10-29 19:25):
#paper Understanding the role of individual units in a deep neural network (https://doi.org/10.1073/pnas.1907375117) PNAS, 2020. 作者通过对Place365数据集上训练得到的VGG16网络的神经元激活图进行上采样观察到了深度学习神经网络中的单个神经元所学习到的概念特征,讨论了这些神经元在“场景分类器”以及生成对抗网络中的“生成器”中的作用,最后讨论了这一发现的应用前景。本项工作的主要研究发现: 1、场景分类器中较“浅”层的神经元倾向于学习到“颜色”、“材质”等抽象概念,较“深”层的神经元倾向于学习到“物体”、“零件”等具体概念。 2、部分神经元对场景识别有重要的作用,关闭这些神经元会导致场景识别能力降低,在多个场景识别任务中都发挥重要作用的神经元具有更好的可解释性。 3、GANs中生成器的神经元学习到的特征与辨别器相反,即,“浅”层的神经元倾向于学习具体概念,而较“深”层的神经元倾向于学习到抽象概念。 4、关闭或启动生成器中的部分神经元,会使生成的图片中去除或增添部分场景元素,同时生成器会根据场景的特性在合适的位置生成物体,因此可以通过操纵GANs中的神经元的激活情况来进行场景绘画。
Abstract:
Deep neural networks excel at finding hierarchical representations that solve complex tasks over large datasets. How can we humans understand these learned representations? In this work, we present network dissection, … >>>
Deep neural networks excel at finding hierarchical representations that solve complex tasks over large datasets. How can we humans understand these learned representations? In this work, we present network dissection, an analytic framework to systematically identify the semantics of individual hidden units within image classification and image generation networks. First, we analyze a convolutional neural network (CNN) trained on scene classification and discover units that match a diverse set of object concepts. We find evidence that the network has learned many object classes that play crucial roles in classifying scene classes. Second, we use a similar analytic method to analyze a generative adversarial network (GAN) model trained to generate scenes. By analyzing changes made when small sets of units are activated or deactivated, we find that objects can be added and removed from the output scenes while adapting to the context. Finally, we apply our analytic framework to understanding adversarial attacks and to semantic image editing. <<<
翻译
665.
林海onrush (2022-10-29 13:58):
#paper,Model Evaluation, Model Selection, and Algorithm Selection in Machine Learning , url : https://arxiv.org/abs/1811.12808#, 本论文回顾了用于解决模型评估、模型选择和算法选择三项任务的不同技术,并参考理论和实证研究讨 论了每一项技术的主要优势和劣势。进而,给出建议以促进机器学习研究与应用方面的最佳实践。 详细论文解析见下面pdf
Abstract:
The correct use of model evaluation, model selection, and algorithm selection techniques is vital in academic machine learning research as well as in many industrial settings. This article reviews different … >>>
The correct use of model evaluation, model selection, and algorithm selection techniques is vital in academic machine learning research as well as in many industrial settings. This article reviews different techniques that can be used for each of these three subtasks and discusses the main advantages and disadvantages of each technique with references to theoretical and empirical studies. Further, recommendations are given to encourage best yet feasible practices in research and applications of machine learning. Common methods such as the holdout method for model evaluation and selection are covered, which are not recommended when working with small datasets. Different flavors of the bootstrap technique are introduced for estimating the uncertainty of performance estimates, as an alternative to confidence intervals via normal approximation if bootstrapping is computationally feasible. Common cross-validation techniques such as leave-one-out cross-validation and k-fold cross-validation are reviewed, the bias-variance trade-off for choosing k is discussed, and practical tips for the optimal choice of k are given based on empirical evidence. Different statistical tests for algorithm comparisons are presented, and strategies for dealing with multiple comparisons such as omnibus tests and multiple-comparison corrections are discussed. Finally, alternative methods for algorithm selection, such as the combined F-test 5x2 cross-validation and nested cross-validation, are recommended for comparing machine learning algorithms when datasets are small. <<<
翻译
666.
林海onrush (2022-10-29 13:51):
#paper,Formal Algorithms for Transformers,url:https://arxiv.org/pdf/2207.09238.pdf,在过去5年多的时间里,Transfermers在多个领域表现出惊人的效果。但是,对于Transformers算法的描述基本都集中在使用图形、文字描述、或针对优化部分的解释,并没有一篇论文给出一个较为完整的Algorithm伪代码。deepmind官方给出了形式化算法伪代码,论文详解见下面PDF
Abstract:
This document aims to be a self-contained, mathematically precise overview of transformer architectures and algorithms (*not* results). It covers what transformers are, how they are trained, what they are used … >>>
This document aims to be a self-contained, mathematically precise overview of transformer architectures and algorithms (*not* results). It covers what transformers are, how they are trained, what they are used for, their key architectural components, and a preview of the most prominent models. The reader is assumed to be familiar with basic ML terminology and simpler neural network architectures such as MLPs. <<<
翻译
667.
林海onrush (2022-10-29 13:25):
#paper,CAUSAL DISCOVERY WITH REINFORCEMENT LEARNING,论文地址:https://arxiv.org/pdf/1906.04477.pdf,官方视频介绍:https://iclr.cc/virtual_2020/poster_S1g2skStPB.html, 因果研究作为下一个潜在的热点,已经吸引了机器学习/深度学习领域的的广泛关注,因果研究中一个经典的问题是「因果发现」问题——从被动可观测的数据中发现潜在的因果图结构。 此论文是华为诺亚方舟实验室被 ICLR 2020 接收的一篇满分论文。在此论文中,华为诺亚方舟实验室因果研究团队将强化学习应用到打分法的因果发现算法中,通过基于自注意力机制的 encoder-decoder 神经网络模型探索数据之间的关系,结合因果结构的条件,并使用策略梯度的强化学习算法对神经网络参数进行训练,最终得到因果图结构。在学术界常用的一些数据模型中,该方法在中等规模的图上的表现优于其他方法,包括传统的因果发现算法和近期的基于梯度的算法。同时该方法非常灵活,可以和任意的打分函数结合使用。
Abstract:
Discovering causal structure among a set of variables is a fundamental problem in many empirical sciences. Traditional score-based casual discovery methods rely on various local heuristics to search for a … >>>
Discovering causal structure among a set of variables is a fundamental problem in many empirical sciences. Traditional score-based casual discovery methods rely on various local heuristics to search for a Directed Acyclic Graph (DAG) according to a predefined score function. While these methods, e.g., greedy equivalence search, may have attractive results with infinite samples and certain model assumptions, they are usually less satisfactory in practice due to finite data and possible violation of assumptions. Motivated by recent advances in neural combinatorial optimization, we propose to use Reinforcement Learning (RL) to search for the DAG with the best scoring. Our encoder-decoder model takes observable data as input and generates graph adjacency matrices that are used to compute rewards. The reward incorporates both the predefined score function and two penalty terms for enforcing acyclicity. In contrast with typical RL applications where the goal is to learn a policy, we use RL as a search strategy and our final output would be the graph, among all graphs generated during training, that achieves the best reward. We conduct experiments on both synthetic and real datasets, and show that the proposed approach not only has an improved search ability but also allows a flexible score function under the acyclicity constraint. <<<
翻译
668.
尹志 (2022-10-27 20:44):
#paper doi: https://doi.org/10.48550/arXiv.1708.02002,Focal Loss for Dense Object Detection. (ICCV 2017) 这是一篇目标检测领域的经典的论文,我们知道,一直以来,目标检测领域有两类模型,单阶段和二阶段检测模型。前者以yolo和ssd为主,后者基本上是R-CNN派生出来的。一般而言,单阶段的目标检测算法速度快于二阶段检测算法,而准确性上弱于二阶段算法。原理上,二阶段检测算法基本是第一步生成一堆的候选目标框,然后第二步精准分类这些候选目标框;而单阶段检测算法是直接生成一堆(大量)的检测框。那么是不是提出一个单阶段的检测算法,速度也快,准确性也可以媲美二阶段算法呢?文章认为,单阶段在准确性上目前比不过二阶段算法的原因,是因为存在类别不平衡的问题。在二阶段算法中,我们通过第一阶段已经过滤了大多数的背景样本了,但单阶段算法一次生成的候选框非常密集,其中前景-背景类别的不平衡就非常严重,这也导致准确率上不去。因此作者提出,我们在常规的交叉熵里引入一个缩放因子,这个缩放因子在训练中能够自动对容易的样本进行降权重,从而让模型能更好的处理难例。这就是大名鼎鼎的focal loss。基于focal loss,作者设计了一个单阶段目标检测网络:RetinaNet, 通过实验对比,RetinaNet不论在速度上还是准确性上,都获得了SOTA的性能,在COCO数据集上获得了39.1的AP(这在当年是非常优秀的成绩)
Abstract:
The highest accuracy object detectors to date are based on a two-stage approach popularized by R-CNN, where a classifier is applied to a sparse set of candidate object locations. In … >>>
The highest accuracy object detectors to date are based on a two-stage approach popularized by R-CNN, where a classifier is applied to a sparse set of candidate object locations. In contrast, one-stage detectors that are applied over a regular, dense sampling of possible object locations have the potential to be faster and simpler, but have trailed the accuracy of two-stage detectors thus far. In this paper, we investigate why this is the case. We discover that the extreme foreground-background class imbalance encountered during training of dense detectors is the central cause. We propose to address this class imbalance by reshaping the standard cross entropy loss such that it down-weights the loss assigned to well-classified examples. Our novel Focal Loss focuses training on a sparse set of hard examples and prevents the vast number of easy negatives from overwhelming the detector during training. To evaluate the effectiveness of our loss, we design and train a simple dense detector we call RetinaNet. Our results show that when trained with the focal loss, RetinaNet is able to match the speed of previous one-stage detectors while surpassing the accuracy of all existing state-of-the-art two-stage detectors. Code is at: this https URL. <<<
翻译
669.
白鸟 (2022-10-27 09:36):
#paper doi:#paper doi:https://doi.org/10.1038/s41587-022-01468-y Haplotype-aware analysis of somatic copy number variations from single-cell transcriptomes. 单细胞转录组体细胞拷贝数变异的单倍型感知分析 基因组不稳定性和转录程序的异常改变都在癌症中发挥重要作用。单细胞 RNA 测序 (scRNA-seq) 在一次检测中能够同时研究肿瘤异质性的遗传和非遗传来源。虽然有许多工具可以从外显子组和全基因组测序数据中识别CNV,针对单细胞RNA-seq数据中检测CNV的方法非常稀缺。常用的inferCNV和copyKAT都只是利用转录组的基因表达信息进行CNV推断。最近,哈佛医学院的研究者提出了一种计算方法,Numbat,它将基于群体的定相(population-based phasing)获得的单倍型信息与等位基因和表达信号相结合,能准确推断单个细胞中的等位基因特异性CNV并重建它们的谱系关系。也就是说它通过基因表达和等位基因两个证据链,进行联合推断,避免CNV推断误判。Numbat利用亚克隆之间的进化关系来迭代推断单细胞拷贝数分布和肿瘤克隆系统发育。比其他工具进行基准测试,对包括多发性骨髓瘤、胃癌、乳腺癌和甲状腺癌在内的 22 个肿瘤样本的分析表明,Numbat可以重建肿瘤拷贝数分布,并准确识别肿瘤微环境中的恶性细胞。Numbat 不需要样本匹配的 DNA 数据,也不需要先验基因分型,适用于广泛的实验环境和癌症类型。总之,Numbat 可以扩展单细胞RNA-seq数据来探测细胞的CNV景观以及转录组景观。需要思考的是我们可能需要更多不同遗传背景的人群定相单倍型信息来辅助推断。另外,肿瘤基线倍性估计仍是拷贝数分析中的有挑战性的问题。
IF:33.100Q1 Nature biotechnology, 2023-03. DOI: 10.1038/s41587-022-01468-y PMID: 36163550
Abstract:
Genome instability and aberrant alterations of transcriptional programs both play important roles in cancer. Single-cell RNA sequencing (scRNA-seq) has the potential to investigate both genetic and nongenetic sources of tumor … >>>
Genome instability and aberrant alterations of transcriptional programs both play important roles in cancer. Single-cell RNA sequencing (scRNA-seq) has the potential to investigate both genetic and nongenetic sources of tumor heterogeneity in a single assay. Here we present a computational method, Numbat, that integrates haplotype information obtained from population-based phasing with allele and expression signals to enhance detection of copy number variations from scRNA-seq. Numbat exploits the evolutionary relationships between subclones to iteratively infer single-cell copy number profiles and tumor clonal phylogeny. Analysis of 22 tumor samples, including multiple myeloma, gastric, breast and thyroid cancers, shows that Numbat can reconstruct the tumor copy number profile and precisely identify malignant cells in the tumor microenvironment. We identify genetic subpopulations with transcriptional signatures relevant to tumor progression and therapy resistance. Numbat requires neither sample-matched DNA data nor a priori genotyping, and is applicable to a wide range of experimental settings and cancer types. <<<
翻译
670.
哪有情可长 (2022-10-26 20:33):
#paper doi:#paper doi:doi:10.1038/ng.592, OsSPL14 promotes panicle branching and higher grainproductivity in rice.作物产量三要素主要是亩穗数、穗粒数和千粒重。提高产量主要是提高三要素之间的协同作用。水稻的第二次“绿色革命”是通过降低株高来增加了水稻的产量。而现在有人认为在水稻中IPA这个基因是新型的"绿色革命"基因。该基因能够能够在水稻的生殖期通过在水稻幼穗内高表达促进水稻穗分枝和籽粒产量。小麦中关于SBP蛋白的研究有很多,通过同源blast,在小麦中也鉴定到了IPA的同源基因。文章在2017年发表在《Plant Physiology》“Functional conservation and divergence among homoeologs of TaSPL20 and TaSPL21, two SBP-box genes governing yield-related traits in hexaploid wheat”,作者发现普通小麦部分同源基因TaSPL20和TaSPL21在小麦长期的驯化和遗传改良过程中产生功能分化,其优异的自然变异在我国小麦育种进程中受到了定向选择并被广泛应用,但是这个文中中验证基因由于当初小麦转基因比较难,文中中验证是在水稻中进行的,证明该基因增加了籽粒大小,从而增加了产量。如果是在小麦中验证会更好。
IF:31.700Q1 Nature genetics, 2010-Jun. DOI: 10.1038/ng.592 PMID: 20495564
Abstract:
Identification of alleles that improve crop production and lead to higher-yielding varieties are needed for food security. Here we show that the quantitative trait locus WFP (WEALTHY FARMER'S PANICLE) encodes … >>>
Identification of alleles that improve crop production and lead to higher-yielding varieties are needed for food security. Here we show that the quantitative trait locus WFP (WEALTHY FARMER'S PANICLE) encodes OsSPL14 (SQUAMOSA PROMOTER BINDING PROTEIN-LIKE 14, also known as IPA1). Higher expression of OsSPL14 in the reproductive stage promotes panicle branching and higher grain yield in rice. OsSPL14 controls shoot branching in the vegetative stage and is affected by microRNA excision. We also demonstrate the feasibility of using the OsSLP14(WFP) allele to increase rice crop yield. Introduction of the high-yielding OsSPL14(WFP) allele into the standard rice variety Nipponbare resulted in increased rice production. <<<
翻译
671.
周周复始 (2022-10-26 20:17):
#paper doi: https://doi.org/10.1101/2021.03.04.433968,Deep Diffusion MRI Registration (DDMReg): A Deep Learning Method for Diffusion MRI Registration。本文基于深度学习提出了新的配准框架,用于dmri数据的配准。由于dmri数据既包含水分子扩散强度也包含水扩散方向信息,所以配准dmri,既要使全脑解剖结构对齐也要让纤维束方向保持一致,传统配准方法存在的问题是要么不包含方向信息,要么是专门针对纤维束进行配准不能保证全脑结构的对齐。本文方法的输入数据包含了代表全脑解剖结构信息的FA图像和代表纤维束方向的TOM图像,通过一个基于voxelmorph改进后的DDMReg网络架构,训练出的模型效果与最先进的四种方法(SyN,DTI-Tk,MRReg,voxelmorph)相比是最优的。
Abstract:
AbstractIn this paper, we present a deep learning method, DDMReg, for accurate registration between diffusion MRI (dMRI) datasets. In dMRI registration, the goal is to spatially align brain anatomical structures … >>>
AbstractIn this paper, we present a deep learning method, DDMReg, for accurate registration between diffusion MRI (dMRI) datasets. In dMRI registration, the goal is to spatially align brain anatomical structures while ensuring that local fiber orientations remain consistent with the underlying white matter fiber tract anatomy. DDMReg is a novel method that uses joint whole-brain and tract-specific information for dMRI registration. Based on the successful VoxelMorph framework for image registration, we propose a novel registration architecture that leverages not only whole brain information but also tract-specific fiber orientation information. DDMReg is an unsupervised method for deformable registration between pairs of dMRI datasets: it does not require nonlinearly pre-registered training data or the corresponding deformation fields as ground truth. We perform comparisons with four state-of-the-art registration methods on multiple independently acquired datasets from different populations (including teenagers, young and elderly adults) and different imaging protocols and scanners. We evaluate the registration performance by assessing the ability to align anatomically corresponding brain structures and ensure fiber spatial agreement between different subjects after registration. Experimental results show that DDMReg obtains significantly improved registration performance compared to the state-of-the-art methods. Importantly, we demonstrate successful generalization of DDMReg to dMRI data from different populations with varying ages and acquired using different acquisition protocols and different scanners. <<<
翻译
672.
庞庞 (2022-10-26 15:51):
#paper doi: https://doi.org/10.1016/j.neuron.2022.08.012,Automated optimization of TMS coil placement for personalized functional network engagement。本文提出了基于个体化的功能网络分区的TMS刺激方法(TANS)。TMS用于治疗多种精神和神经系统疾病,但由于个体的功能网络分布不同,所以个体反应是高度可变的。传统的基于组水平的TMS方法可能会无意中针对抑郁症患者的不同功能网络,导致疗效不佳。而作者开发的 TANS方法,选择的线圈位置则是使得个体目标功能网络受刺激占比最高的位置。作者从计算机和活体验证两方面,验证了TASN可以提高被试的刺激特异性。
IF:14.700Q1 Neuron, 2022-10-19. DOI: 10.1016/j.neuron.2022.08.012 PMID: 36113473
Abstract:
Transcranial magnetic stimulation (TMS) is used to treat multiple psychiatric and neurological conditions by manipulating activity in particular brain networks and circuits, but individual responses are highly variable. In clinical … >>>
Transcranial magnetic stimulation (TMS) is used to treat multiple psychiatric and neurological conditions by manipulating activity in particular brain networks and circuits, but individual responses are highly variable. In clinical settings, TMS coil placement is typically based on either group average functional maps or scalp heuristics. Here, we found that this approach can inadvertently target different functional networks in depressed patients due to variability in their functional brain organization. More precise TMS targeting should be feasible by accounting for each patient's unique functional neuroanatomy. To this end, we developed a targeting approach, termed targeted functional network stimulation (TANS). The TANS approach improved stimulation specificity in silico in 8 highly sampled patients with depression and 6 healthy individuals and in vivo when targeting somatomotor functional networks representing the upper and lower limbs. Code for implementing TANS and an example dataset are provided as a resource. <<<
翻译
673.
王昊 (2022-10-25 10:11):
#paper doi: 10.48550/arXiv.2110.07342 So Yeon Min, Devendra Singh Chaplot, Pradeep Ravikumar, Yonatan Bisk, and Ruslan Salakhutdinov. 2022. FILM: Following Instructions in Language with Modular Methods. Retrieved July 13, 2022 from http://arxiv.org/abs/2110.07342. 应用于视觉语言导航任务的算法文章,目前在ALFRED数据集下排名第4的方法。本文提出了一种具有结构化表示的模块化方法,(1)构建场景的语义地图,(2)使用语义搜索策略进行探索,以实现自然语言目标。Film的四个组件:1.将语言指令转换成结构化形式(语言处理)2.将以自我为中心的视觉输入转换为语义度量图(语义映射)3. 将以自我为中心的视觉输入转换为语义度量图(语义搜索策略)4. 输出后续导航/交互操作(确定性策略)。FILM不需要任何提供顺序指导的输入,即专家轨迹或低级语言指令(用来指导顺序)。
Abstract:
Recent methods for embodied instruction following are typically trained end-to-end using imitation learning. This often requires the use of expert trajectories and low-level language instructions. Such approaches assume that neural … >>>
Recent methods for embodied instruction following are typically trained end-to-end using imitation learning. This often requires the use of expert trajectories and low-level language instructions. Such approaches assume that neural states will integrate multimodal semantics to perform state tracking, building spatial memory, exploration, and long-term planning. In contrast, we propose a modular method with structured representations that (1) builds a semantic map of the scene and (2) performs exploration with a semantic search policy, to achieve the natural language goal. Our modular method achieves SOTA performance (24.46 %) with a substantial (8.17 % absolute) gap from previous work while using less data by eschewing both expert trajectories and low-level instructions. Leveraging low-level language, however, can further increase our performance (26.49 %). Our findings suggest that an explicit spatial memory and a semantic search policy can provide a stronger and more general representation for state-tracking and guidance, even in the absence of expert trajectories or low-level instructions. <<<
翻译
674.
徐炳祥 (2022-10-23 15:49):
#paper doi: 10.1016/j.cell.2022.09.006 Cell, 2022, Repression and 3D-restructuring resolves regulatory conflicts in evolutionarily rearranged genomes。本文通过研究位于同一个增强子作用域(也是同一个TAD)内的两个基因Zfp42和Fat1在胚胎发育中的表达模式,发现他们各自受该区域内特定增强子的影响而互不干扰的独立调控,从而指出存在一种可在不改变基因组空间构象和增强子作用域的前提下屏蔽增强子对特定基因的作用的机制。进一步,他们通过分析DNA甲基转移酶敲除对两个基因启动子区域甲基化水平的影响和相应的表达图谱的变化指出DNA甲基化可能是此类机制中的一种。最后,通过基因共表达分析,作者指出,此种在同一个增强子作用域内出现的基因表达调控模式的多变性多见于发育相关基因而少见于持家基因,且可被DNA甲基化介导的转录抑制所解释。
IF:45.500Q1 Cell, 2022-09-29. DOI: 10.1016/j.cell.2022.09.006 PMID: 36179666
Abstract:
Regulatory landscapes drive complex developmental gene expression, but it remains unclear how their integrity is maintained when incorporating novel genes and functions during evolution. Here, we investigated how a placental … >>>
Regulatory landscapes drive complex developmental gene expression, but it remains unclear how their integrity is maintained when incorporating novel genes and functions during evolution. Here, we investigated how a placental mammal-specific gene, Zfp42, emerged in an ancient vertebrate topologically associated domain (TAD) without adopting or disrupting the conserved expression of its gene, Fat1. In ESCs, physical TAD partitioning separates Zfp42 and Fat1 with distinct local enhancers that drive their independent expression. This separation is driven by chromatin activity and not CTCF/cohesin. In contrast, in embryonic limbs, inactive Zfp42 shares Fat1's intact TAD without responding to active Fat1 enhancers. However, neither Fat1 enhancer-incompatibility nor nuclear envelope-attachment account for Zfp42's unresponsiveness. Rather, Zfp42's promoter is rendered inert to enhancers by context-dependent DNA methylation. Thus, diverse mechanisms enabled the integration of independent Zfp42 regulation in the Fat1 locus. Critically, such regulatory complexity appears common in evolution as, genome wide, most TADs contain multiple independently expressed genes. <<<
翻译
675.
DeDe宝 (2022-10-23 15:17):
#paper The source of systematic errors in human path integration,DOI: 10.7939/r3-ftnc(preprint)这篇文献使用交叉验证模型(cross-validation modeling)比较 M1(Encoding-error model)M2(execution-error model)M3(bi-component model)三种模型对人类在路径整合中的系统误差(压缩模式)的拟合和解释。结果支持了bi-component model的假设,该模型同时考虑了编码过程和执行过程引入的系统误差。此外,使用单响应输入条件无法将双成分模型与编码误差模型和执行误差模型分离,表明使用多个出站路径的多个入站响应进行交叉验证建模可能是了解人类路径整合的强大工具。
Abstract:
Triangle completion is a task widely used to study human path integration, an important navigation method relying on idiothetic cues. Systematic biases (compression patterns in the inbound responses) have been … >>>
Triangle completion is a task widely used to study human path integration, an important navigation method relying on idiothetic cues. Systematic biases (compression patterns in the inbound responses) have been well documented in human triangle completion. However, the sources of systematic biases remain controversial. We used cross-validation modeling to compare three plausible theoretical models that assume that systematic errors occur in the encoding outbound path solely (encoding-error model), executing the inbound responses solely (execution-error model), and both (bi-component model), respectively. The data for cross-validation modeling are from a previous study (Qi et al., 2021), in which participants learned three objects’ locations (one at the path origin, that is, home) very well before walking each outbound path and then pointed to the objects’ original locations after walking the outbound path. The modeling algorithm used one inbound response (i.e., response to the home) or multiple inbound responses (i.e., responses to two non-home locations and the home) for each outbound path. The algorithm of using multiple inbound responses demonstrated that the bi-component model outperformed the other models in accounting for the systematic errors. This finding suggests that both encoding the outbound path and executing the inbound responses contribute to the systematic biases in human path integration. In addition, the results showed that the algorithm using only the home response could not distinguish among these three models, suggesting that the typical triangle-completion task with only the home response for each outbound path cannot determine the sources of the systematic biases. <<<
翻译
676.
张浩彬 (2022-10-20 16:20):
#paper 1.Unsupervised Scalable Representation Learning for Multivariate Time Series,https://doi.org/10.48550/arXiv.1901.10738 论文关键是:正负样本构造, triplet loss以及因果空洞卷积 适用:该无监督学习模型可以用于不定长的序列;短序列及长序列均可使用; 代码:https://github.com/White-Link/UnsupervisedScalableRepresentationLearningTimeSeries 正负样本构造: 有N个序列对于某序列,随机选择长度,构造一个子序列ref。在这个子序列中,随机抽样一个子序列作为正样本pos;从其他序列(如果有的话)中随机抽样K个作为负样本neg;其中K是超参数 编码器有三个要求:(1)能够提取序列特征;(2)允许变长输入;(3)可以节省时间和内存;(个人觉得,只是为了给使用卷积找的理由);因此使用exponentially dilated causal convolutions作为特征提取器代替传统的rnn、lstm 改造的triplet loss 在时间序列分类任务中结果表明由于现有的无监督方法,并且不差于有监督方法。在序列预测任务中,没做太多的比较 在单序列分类任务:使用了UCR数据集上的所有时间序列分类任务
Abstract:
Time series constitute a challenging data type for machine learning algorithms, due to their highly variable lengths and sparse labeling in practice. In this paper, we tackle this challenge by … >>>
Time series constitute a challenging data type for machine learning algorithms, due to their highly variable lengths and sparse labeling in practice. In this paper, we tackle this challenge by proposing an unsupervised method to learn universal embeddings of time series. Unlike previous works, it is scalable with respect to their length and we demonstrate the quality, transferability and practicability of the learned representations with thorough experiments and comparisons. To this end, we combine an encoder based on causal dilated convolutions with a novel triplet loss employing time-based negative sampling, obtaining general-purpose representations for variable length and multivariate time series. <<<
翻译
677.
张德祥 (2022-10-18 10:58):
#paper https://doi.org/10.48550/arXiv.2208.10601Deriving time-averaged active inference from control principles 通过观察随时反馈调整规划的理论实现, 假设固定的动作空间和前馈规划,这可能导致非常高维的递归优化问题。这些假设在经验上和计算上都是有问题的。有机体并不是生来就知道[9];他们学习[40]. 噪音[13,32], 不确定[23], 和可变性[47] 在运动控制方面不够完善,因此必须通过在线反馈来稳定运动控制。 随机最优反馈控制需要一个最优性原则,允许在行动步骤之间整合观察。而不是递归优化单独的动作,通过观察随时反馈调整规划序列。 尽管优化了“全局”(不确定)惊奇率(等式),它只需要在情境中规划和调整行为。 泰德帕里和 Ok[55] 1998 年发表了第一个基于模型的 RL 算法,而 Baxter 和 Bartlett[5] 给出了有偏的政策梯度估计量。亚历山大和布朗又花了十年时间[2]以给出平均成本时间差异学习的递归分解。张与罗斯[61] 直到最近,我才首次发表了“深度”强化学习算法(基于函数逼近)对平均成本标准的适应,该标准仍然是无模型的。Jafarnia-Jahromi 等人[26]最近给出了第一个算法 , 用 于 求 解 具 有 已 知 观 测 密 度 和 未 知 动 态 的 无 限 时 域 平 均 代 价 部 分 可 观 测 问 题 。 结论 这结束了主动推理的无限视野、平均惊奇公式的推导。由于我们的公式将行为情节置于情境中,所以尽管优化了“全局”(不确定)惊奇率(等式),它只需要在情境中规划和调整行为(例如,从时间步长 1 到 T)15). 我们认为,这种积极推理公式可以推进基于模型的概率方法,分层反馈控制[40,33].
Abstract:
Active inference offers a principled account of behavior as minimizing average sensory surprise over time. Applications of active inference to control problems have heretofore tended to focus on finite-horizon or … >>>
Active inference offers a principled account of behavior as minimizing average sensory surprise over time. Applications of active inference to control problems have heretofore tended to focus on finite-horizon or discounted-surprise problems, despite deriving from the infinite-horizon, average-surprise imperative of the free-energy principle. Here we derive an infinite-horizon, average-surprise formulation of active inference from optimal control principles. Our formulation returns to the roots of active inference in neuroanatomy and neurophysiology, formally reconnecting active inference to optimal feedback control. Our formulation provides a unified objective functional for sensorimotor control and allows for reference states to vary over time. <<<
翻译
678.
张德祥 (2022-10-17 20:49):
#paper https://doi.org/10.1016/j.biopsycho.2021.108242 Interoception as modeling, allostasis as control 大脑首先要维持身体的正常状态,还要对将要到来的未来需要提前准备,这需要大脑对身体有建模,大脑对外部世界进行建模,对自身身体也有建模,有关于自我身体的模型,并控制及预测未来身体的需求,比如比赛前的预热、深呼吸。包括管理分泌系统,免疫系统,消化系统等。 心理学家用许多术语来指代内部模型,包括interoception,包括记忆,信念,知觉推理,无意识推理,具身模拟,概念和类别,受控幻觉,预测。 大脑正在预测性地调节身体,这是一个运动控制的问题,而不是感知世界的问题。这是一个沿着期望的轨迹调节身体以实现效率的问题 对身体的调控分两方面,一方面如果提升营养供应,另一方面就要提升废物代谢,这种成对的控制几乎出现在全身的各种调节模式中。 具身决策包括所有三种形式的不确定性,这三种形式的不确定性都受制于非稳态调节:关于生理有效的不确定性,关于运动结果的不确定性,以及关于外部世界的不确定性。 文章提出了非稳态路径积分控制(APIC) Allostatic Path-Integral Control (APIC) 模型。APIC 有一个简单的核心思想:就像知觉概念是身体感觉表面的内部模型一样 15,92,14],行动概念也作为潜在行为及其预测结果的内部模型。
Abstract:
The brain regulates the body by anticipating its needs and attempting to meet them before they arise - a process called allostasis. Allostasis requires a model of the changing sensory … >>>
The brain regulates the body by anticipating its needs and attempting to meet them before they arise - a process called allostasis. Allostasis requires a model of the changing sensory conditions within the body, a process called interoception. In this paper, we examine how interoception may provide performance feedback for allostasis. We suggest studying allostasis in terms of control theory, reviewing control theory's applications to related issues in physiology, motor control, and decision making. We synthesize these by relating them to the important properties of allostatic regulation as a control problem. We then sketch a novel formalism for how the brain might perform allostatic control of the viscera by analogy to skeletomotor control, including a mathematical view on how interoception acts as performance feedback for allostasis. Finally, we suggest ways to test implications of our hypotheses. <<<
翻译
679.
张德祥 (2022-10-11 09:40):
#paper DOI: https://doi.org/10.1145/3428208 Scaling Exact Inference for Discrete Probabilistic Programs 概率推理的计算挑战是应用概率编程的主要障碍,如何解决?如何利用程序的结构来分解推理,如何解耦分不的结构和参数?如何证明编译的语义正确? dice 语言使用weighted model counting (WMC)推理,使用weighted Boolean formulas (WBF) 将代码编译为 binary decision diagrams (BDDs) to represent these formulas; experiments in Section 5 show Dice performing exact inference on a real-world probabilistic program that is 1.9MB large. 由于避免了指数爆炸,dice 编译的大小是线性的,计算是有保证的,编译方法是有数学证明理论的保证。 dice跟之前概率编程很大的不同是,同时支持常规编程语言的结构 if else for等。 一个关键挑战是dice支持任意观察,dice编译程序到两种BDD,一个支持程序的任意执行忽略观察,另一个表示满足观察的所有执行。 dice 开源。 The key insight is to separate the logical representation of the state space of the program from the probabilities 一旦程序被编译成 BDD,Dice 通过 WMC 对原始概率程序进行推理。至关重要的是,它这样做并没有穷尽列举所有的路径或模型。(高效) 通过条件独立进行抽象降低计算复杂度。(独立性,条件独立,局部结构) 补充参考: https://mp.weixin.qq.com/s/Rks2VGLz8G9XS3IGR7xegw
Abstract:
Probabilistic programming languages (PPLs) are an expressive means of representing and reasoning about probabilistic models. The computational challenge of probabilistic inference remains the primary roadblock for applying PPLs in practice. … >>>
Probabilistic programming languages (PPLs) are an expressive means of representing and reasoning about probabilistic models. The computational challenge of probabilistic inference remains the primary roadblock for applying PPLs in practice. Inference is fundamentally hard, so there is no one-size-fits all solution. In this work, we target scalable inference for an important class of probabilistic programs: those whose probability distributions are discrete . Discrete distributions are common in many fields, including text analysis, network verification, artificial intelligence, and graph analysis, but they prove to be challenging for existing PPLs. We develop a domain-specific probabilistic programming language called Dice that features a new approach to exact discrete probabilistic program inference. Dice exploits program structure in order to factorize inference, enabling us to perform exact inference on probabilistic programs with hundreds of thousands of random variables. Our key technical contribution is a new reduction from discrete probabilistic programs to weighted model counting (WMC). This reduction separates the structure of the distribution from its parameters, enabling logical reasoning tools to exploit that structure for probabilistic inference. We (1) show how to compositionally reduce Dice inference to WMC, (2) prove this compilation correct with respect to a denotational semantics, (3) empirically demonstrate the performance benefits over prior approaches, and (4) analyze the types of structure that allow Dice to scale to large probabilistic programs. <<<
翻译
680.
笑对人生 (2022-10-08 00:00):
#paper doi: 10.1038/s41523-018-0066-6. Migrating the SNP array-based homologous recombination deficiency measures to next generation sequencing data of breast cancer.  NPJ Breast Cancer.  2018 Jul 2;4:16. 同源重组修复(homologous recombination repair,HRR)是DNA双链断裂(double strand break,DSB)的首选修复方式。同源重组修复缺陷(homologous recombination defificiency,HRD)通常指细胞水平上的HRR功能障碍状态,可由HRR相关基因胚系突变或体细胞突变以及表观遗传失活等诸多因素导致,常存在于多种恶性肿瘤中,其中在卵巢癌、乳腺癌、胰腺导管癌、前列腺癌等肿瘤尤其突出。当HRD存在时,DSB会过度依赖非同源末端连接(non-homologous end joining,NHEJ)、微同源末端连接(microhomology mediated end joining,MMEJ)和单链退火途径(single-strand annealing,SSA)等低保真、高易错的替代性DNA损伤修复途径,从而极可能造成核酸序列的插入/缺失,拷贝数异常,并引起染色体交联,造成基因组和染色体不稳定。HRD临床检验所描述的是肿瘤基因组特定改变,也称为基因组瘢痕(genomic scar)。HRD评分(HRD score)可以用来反映肿瘤样本因HRR通路异常而导致的肿瘤样本基因组不稳定的情况。HRD score计算了三种得分的和:端粒等位基因不平衡(telomeric allelic imbalance,TAI或NtAI)评分,杂合缺失(loss of heterozygosity,LOH)评分和大片段迁移(large-scale state transition)评分。HRD评分的检测可采用SNP芯片或NGS平台。 本研究开发了一个名为scarHRD的软件包。利用scarHRD对SNP芯片和NGS平台(WES或WGS)的数据计算HRD评分,结果发现两个平台之间具有很好的相关性(Pearson相关系数在0.73-0.87之间)。对来自TCGA的三阴性乳腺癌BRCA突变和BRCA野生型队列进行分析,发现与BRCA1/2野生型患者相比,利用scarHRD计算HRD评分在突变型患者中更高,ROC曲线对应的AUC面积达80.8%,表明scarHRD能够成功反映真实的生物学功能。乳腺癌1号基因(breast cancer 1,BRCA1)是抑癌基因,主要参与DNA断裂修复过程。当BRCA发生功能缺失会导致双链断裂的DNA修复不能通过同源重组修复,进而引起基因组不稳定(genomic instability,GI)。
Abstract:
The first genomic scar-based homologous recombination deficiency (HRD) measures were produced using SNP arrays. As array-based technology has been largely replaced by next generation sequencing approaches, it has become important … >>>
The first genomic scar-based homologous recombination deficiency (HRD) measures were produced using SNP arrays. As array-based technology has been largely replaced by next generation sequencing approaches, it has become important to develop algorithms that derive the same type of genomic scar scores from next generation sequencing (whole exome "WXS", whole genome "WGS") data. In order to perform this analysis, we introduce here the scarHRD R package and show that using this method the SNP array-based and next generation sequencing-based derivation of HRD scores show good correlation (Pearson correlation between 0.73 and 0.87 depending on the actual HRD measure) and that the NGS-based HRD scores distinguish similarly well between BRCA mutant and BRCA wild-type cases in a cohort of triple-negative breast cancer patients of the TCGA data set. <<<
翻译
回到顶部