principles and practices of interconnection networks One of the greatest challenges faced by designers of digital systems is optimizing the communication and interconnection between system components. Interconnection networks offer an attractive and economical solution to this communication crisis and are fast becoming pervasive in digital systems. Current trends suggest that this communication bottleneck will be even more problematic when designing future generations of machines. Consequently, the anatomy of an interconnection network router and science of interconnection network design will only grow in importance in the coming years.
2021-08-08 17:11:59 8.45MB networ
1
贝叶斯分类是统计学方法。他们可以预测类成员关系的可能性,如给定样本属于一个特定类的概率。贝叶斯分类主要是基于贝叶斯定理,通过计算给定样本属于一个特定类的概率来对给定样本进行分类。
2021-08-07 12:06:11 871KB 机器学习 朴素贝叶斯
本资源是对抗样本领域中首次提出对抗样本概念并提出使用L-BFGS攻击算法的一篇文章的代码实现,使用的语言是Pytorch语言,文件为Jupyter notebook文件,在电脑环境配置无问题的情况下,可以直接运行此代码文件,内含详细注释。
神经网络概述 OVERVIEW OF NEURAL NETWORKS
2021-08-04 19:05:43 1.69MB 神经网络
A Brief Didactic Theoretical Review on Convolutional Neural Networks, Deep Belief Networks and Stacked Auto-Encoders
2021-08-04 15:05:22 988KB 深度学习
P2P优秀论文_Super-proximity routing in structured peer-to-peer overlay networks
1
2021.8.3_NeurIPS-2018-group-equivariant-capsule-networks-Paper.pptx
2021-08-03 21:06:29 4.28MB CAPSULE
1
三个 bound 不如一个 heuristic,三个 heuristic 不如一个trick
2021-08-03 01:13:17 11.68MB 2012年第二版
1
具有时间编码的监督学习的目的是使神经元尖峰化,以使神经元响应给定的突触输入而发出任意的尖峰序列。 近年来,基于突触可塑性的监督学习算法发展Swift。 作为最有效的监督学习算法之一,远程监督方法(ReSuMe)使用常规的基于对的峰值定时依赖的可塑性规则,该规则取决于突触前和突触后峰值的精确定时。 在本文中,使用了基于三重态的依赖于尖峰时序的可塑性,它是一种强大的突触可塑性规则,其作用超出了经典规则,提出了一种新颖的监督学习算法,称为T-ReSuMe,以提高ReSuMe的性能。 所提出的算法已成功应用于各种尖峰序列的学习任务,其中所需的尖峰序列通过泊松过程进行​​编码。 实验结果表明,与传统的ReSuMe算法相比,T-ReSuMe算法具有更高的学习精度和更少的迭代次数,对于解决复杂的时空模式学习问题是有效的。
2021-07-27 22:43:31 294KB Spiking neural networks; Supervised
1
一个经典的LSTM教程,以图形化方式开始,从RNN开始,逐步引入Cell的思想和各种门的思想。 Humans don’t start their thinking from scratch every second. As you read this essay, you understand each word based on your understanding of previous words. You don’t throw everything away and start thinking from scratch again. Your thoughts have persistence. Traditional neural networks can’t do this, and it seems like a major shortcoming. For example, imagine you want to classify what kind of event is happening at every point in a movie. It’s unclear how a traditional neural network could use its reasoning about previous events in the film to inform later ones. Recurrent neural networks address this issue. They are networks with loops in them, allowing information to persist.
2021-07-25 17:28:19 1.87MB LSTM 循环网络
1