Convolutional

Layer-Dependent Importance Sampling for Training Deep and Large Graph Convolutional Networks

Zou D., Hu Z., Wang Y., Jiang S., Sun Y. and Gu Q. Layer-dependent importance sampling for training deep and large graph convolutional networks. NIPS, ......

FastGCN Fast Learning with Graph Convolutional Networks via Importance Sampling

Chen J., Ma T. and Xiao C. FastGCN: fast learning with graph convolutional networks via importance sampling. ICLR, 2018. 概 一般的 GCN 每层通常需要经过所有的结点的 prop ......

Deeper Insights into Graph Convolutional Networks for Semi-Supervised Learning

Li Q., Han Z. and Wu X. Deeper insights into graph convolutional networks for semi-supervised learning. AAAI, 2018. 概 本文分析了 GCN 的实际上就是一种 Smoothing, 但是 ......

Stochastic Training of Graph Convolutional Networks with Variance Reduction

Chen J., Zhu J. and Song L. Stochastic training of graph convolutional networks with variance reduction. ICML, 2018. 概 我们都知道, GCN 虽然形式简单, 但是对于结点个数非常多的 ......

卷积神经网络(Convolutional Neural Network)

前置芝士: 神经网络 #前言 人脑视觉机理,是指视觉系统的信息处理在可视皮层是分级的,大脑的工作过程是一个不断迭代、不断抽象的过程。视网膜在得到原始信息后,首先经由区域V1初步处理得到边缘和方向特征信息,其次经由区域V2的进一步抽象得到轮廓和形状特征信息,如此迭代地经由更多更高层的抽象最后得到更为精 ......

论文翻译:2022_DNS_1th:Multi-scale temporal frequency convolutional network with axial attention for speech enhancement

论文地址:带轴向注意的多尺度时域频率卷积网络语音增强 论文代码:https://github.com/echocatzh/MTFAA-Net 引用:Zhang G, Yu L, Wang C, et al. Multi-scale temporal frequency convolutional n ......
共36篇  :2/2页 首页上一页2下一页尾页