train

Cluster-GCN An Efficient Algorithm for Training Deep Convolution Networks

Chiang W., Liu X., Si S., Li Y., Bengio S. and Hsieh C. Cluster-GCN: An efficient algorithm for training deep and large graph convolutional networks. ......

Vicuna-13B, an open-source chatbot trained by fine-tuning LLaMA

一、项目背景 We introduce Vicuna-13B, an open-source chatbot trained by fine-tuning LLaMA on user-shared conversations collected from ShareGPT. Preliminary ......

论文解读(VAT)《Virtual Adversarial Training: A Regularization Method for Supervised and Semi-Supervised Learning》

论文信息 论文标题:Virtual Adversarial Training: A Regularization Method for Supervised and Semi-Supervised Learning论文作者:Takeru Miyato, S. Maeda, Masanori Koya ......

论文解读《Do We Need Zero Training Loss After Achieving Zero Training Error?》

论文信息 论文标题:Do We Need Zero Training Loss After Achieving Zero Training Error?论文作者:Takashi Ishida, I. Yamane, Tomoya Sakai, Gang Niu, M. Sugiyama论文来源:20 ......
Training Zero Achieving 论文 After

猛读论文13 |【CVPR 2022 UDA】Unleashing Potential of Unsupervised Pre-Training with Intra-Identity Regularization for Person Re-Identification

动机 解决(1)对比学习管道中的增强通常会扭曲人物图像中的判别线索(2)细粒度的局部特征人物图像尚未得到充分探索。 思路 方法 ......

迁移学习(PAT)《Pairwise Adversarial Training for Unsupervised Class-imbalanced Domain Adaptation》

论文信息 论文标题:Pairwise Adversarial Training for Unsupervised Class-imbalanced Domain Adaptation论文作者:Weili Shi, Ronghang Zhu, Sheng Li论文来源:KDD 2022论文地址:dow ......

Layer-Dependent Importance Sampling for Training Deep and Large Graph Convolutional Networks

Zou D., Hu Z., Wang Y., Jiang S., Sun Y. and Gu Q. Layer-dependent importance sampling for training deep and large graph convolutional networks. NIPS, ......

Train the Tesseract OCR engine[how to do]

Training the Tesseract OCR engine is a complex and time-consuming process that involves several steps. Here is an overview of the process: Prepare you ......
Tesseract engine Train OCR the

ziyi-lstm-train代码

lstm的train代码 def train_lstm(net,lr,train_loader,total_epoch): global_step = 1 optimizer = torch.optim.Adam(net.parameters(), lr=lr) scheduler = lr_sch ......
ziyi-lstm-train 代码 train ziyi lstm

BUPT 2023 Spring Training #9

原题:2021“MINIEYE杯”中国大学生算法设计超级联赛(1) 卡在两道题上,然后就没有然后了 A 对于 $i \in [0,\lceil\frac n2\rceil-1] \cap {\mathbb Z}$,取模时一定可以取到($n \equiv i({\rm mod}\ n-i)$) 对于 ......
Training Spring BUPT 2023

GPT模型: Generative Pre-training 生成式无监督预训练

GPT,GPT-2,GPT-3 论文精读【论文精读】_哔哩哔哩_bilibili ELMo:将上下文当作特征,但是无监督的语料和我们真实的语料还是有区别的,不一定符合我们特定的任务,是一种双向的特征提取。 OpenAI GPT: 通过transformer decoder学习出来一个语言模型,不是固 ......
Pre-training Generative training 模型 GPT

Stochastic Training of Graph Convolutional Networks with Variance Reduction

Chen J., Zhu J. and Song L. Stochastic training of graph convolutional networks with variance reduction. ICML, 2018. 概 我们都知道, GCN 虽然形式简单, 但是对于结点个数非常多的 ......

Generative Pre-trained Transformer(GPT)模型技术初探

一、Transformer模型 2017年,Google在论文 Attention is All you need 中提出了 Transformer 模型,其使用 Self-Attention 结构取代了在 NLP 任务中常用的 RNN 网络结构。相比 RNN 网络结构,其最大的优点是可以并行计算。 ......

HNU2019 Summer Training 3 E. Blurred Pictures

E. Blurred Pictures time limit per test 2 seconds memory limit per test 256 megabytes input standard input output standard output Damon loves to take ......
Pictures Training Blurred Summer 2019

论文解读( FGSM)《Adversarial training methods for semi-supervised text classification》

论文信息 论文标题:Adversarial training methods for semi-supervised text classification论文作者:Taekyung Kim论文来源:ICLR 2017论文地址:download 论文代码:download视屏讲解:click 1 背 ......

【IOI2017】Toy Train(博弈)

题目链接:https://uoj.ac/problem/322 分析 “一个点的出边一旦确定就不能改变”这个条件不好处理。通过网上一些题解的分析,可以把问题修改成: 结点的主人每次可以指定任意一条出边(即使之前已经指定了另外一条)。 A 胜利条件:存在一种策略,无论 B 怎么操作,总能使火车无限次经 ......
Train 2017 IOI Toy

迁移学习《Asymmetric Tri-training for Unsupervised Domain Adaptation》

论文信息 论文标题:Asymmetric Tri-training for Unsupervised Domain Adaptation论文作者:Kuniaki Saito, Y. Ushiku, T. Harada论文来源:27 February 2017——ICML论文地址:download 论 ......

【829】sklearn中train_test_split函数中的random_state有什么用?

参考:sklearn.model_selection中train_test_split的坑 参考:sklearn中train_test_split函数中的random_state有什么用? 对 random_state 设置一个固定的值,可以保证每次得到相同的训练集与测试集! ......

「解题报告」ARC123E Training

挺有趣的题,为数不多的自己能切的题。 题意无非就是要你求 $i \in [1, n]$,有多少满足 $a + \lfloor\frac{i}{b} \rfloor = c + \lfloor\frac{i}{d}\rfloor$。 首先移项,得 $a - c = \lfloor\frac{i}{d} ......
Training 报告 123E ARC 123

KPCA matlab代码,可分train和test。 注释清晰

KPCA matlab代码,可分train和test。 注释清晰YID:7220647215929418 ......
注释 代码 matlab train KPCA

Going Deeper With Directly-Trained Larger Spiking Neural Networks

郑重声明:原文参见标题,如有侵权,请联系作者,将会撤销发布! The Thirty-Fifth AAAI Conference on Artificial Intelligence (AAAI-21) Abstract 脉冲神经网络(SNN)在时空信息和事件驱动信号处理的生物可编程编码中很有前途,非 ......

论文解读(Moco v3)《An Empirical Study of Training Self-Supervised Vision Transformers》

论文信息 论文标题:Improved Baselines with Momentum Contrastive Learning论文作者:Xinlei Chen, Saining Xie, Kaiming He论文来源:2021 ICCV论文地址:download 论文代码:download引用次数: ......

迁移学习(IIMT)——《Improve Unsupervised Domain Adaptation with Mixup Training》

论文信息 论文标题:Improve Unsupervised Domain Adaptation with Mixup Training论文作者:Shen Yan, Huan Song, Nanxiang Li, Lincan Zou, Liu Ren论文来源:arxiv 2020论文地址:down ......
共113篇  :4/4页 首页上一页4下一页尾页