首页 > 其他分享 >DeeperGCN All You Need to Train Deeper GCNs

DeeperGCN All You Need to Train Deeper GCNs

时间:2023-05-13 16:33:46浏览次数:43  
标签:mathbf bm beta Train GCNs Need mathcal zeta vu

目录

Li G., Xiong C., Thabet A. and Ghanem B. DeeperGcn: all you need to train deeper gcns. arXiv preprint arXiv:2006.07739

本文介绍了一种连续可微的 aggregation function (极限状态下能够 cover 常用的 mean, max).

符号说明

  • \(\mathcal{G} = (\mathcal{V}, \mathcal{E})\), 图;
  • \(\mathbf{h}_v \in \mathbb{R}^D\), node features;
  • \(\mathbf{h}_e \in \mathbb{R}^C\), edge features;
  • \(\mathcal{N}(v)\), node \(v\) 的一阶邻居.

广义的 aggregation function

  • 一般的 GCN 每一层需要历经:

    \[\mathbf{m}_{vu}^{(l)} = \bm{\rho}^{(l)} (\mathbf{h}_v^{(l)}, \mathbf{h}_v^{(l)}, \mathbf{h}_{e_{vu}}^{(l)}), \: u \in \mathcal{N}(v), \\ \mathbf{m}_v^{(l)} = \bm{\zeta}^{(l)}(\{\mathbf{m}_{vu}^{(l)}| u \in \mathcal{N}(v)\}), \\ \mathbf{h}_v^{(l+1)} = \bm{\phi}^{(l)} (\mathbf{h}_v^{(l)}, \mathbf{m}_v^{(l)}). \]

    其中 \(\bm{\rho}, \bm{\zeta}, \bm{\phi}\) 为可训练的函数. 这里我们主要关注 \(\bm{\zeta}\).

  • SoftMax Aggregation:

    \[\bm{\zeta}_{\beta}(\{\mathbf{m}_{vu}^{(l)}| u \in \mathcal{N}(v)\}) = \sum_{u \in \mathcal{N}_v} \frac{\exp(\beta \mathbf{m}_{vu})}{\sum_{i \in \mathcal{N}(v)} \exp(\beta \mathbf{m}_{vi})} \cdot \mathbf{m}_{vu}. \]

    显然当 \(\beta \rightarrow 0\) 的时候它退化为 \(\text{Mean}(\cdot)\), 当 \(\beta \rightarrow +\infty\) 的时候退化为 \(\text{Max}(\cdot)\).

  • PowerMean Aggregation:

    \[\bm{\zeta}_{p}(\{\mathbf{m}_{vu}^{(l)}| u \in \mathcal{N}(v)\}) =(\frac{1}{|\mathcal{N}_v|} \sum_{u \in \mathcal{N}_v} \mathbf{m}_{vu}^p)^{1/p}, \]

    当 \(p = 1\) 的时候, 退化为 \(\text{Mean}(\cdot)\), 当 \(p \rightarrow +\infty\) 的时候退化为 \(\text{Max}(\cdot)\), 当 \(p=-1, p \rightarrow 0\) 的时候分别退化为 harmonic 和 geometric mean aggregation.

代码

official

标签:mathbf,bm,beta,Train,GCNs,Need,mathcal,zeta,vu
From: https://www.cnblogs.com/MTandHJ/p/17397616.html

相关文章

  • Fine-tuning Pretrained Network技术研究
    一、基本概念定义(1)数据域(domain)用D={χ,P(X)}表示,它包含两部分:特征空间χ和边缘概率分布P(X)其中X={x1,...xn}∈χ在文本分类任务中,把每一个单词看作二值的特征即出现或者不出现,所有检索词向量的空间就是χ,xi对应某一文本第i个词向量的值,X就是特定的学习样本。如果说两个数据域不同,......
  • SemiEng20230413-What Designers Need To Know About GAA
    Nanowire与nanosheet争议仍然存在,业界还没确定谁更适合作下一代主流逻辑器件。对任何新器件,第一代都是用来学习试验的,后面再迭代升级。FinFET不能继续缩微的原因:fin之间要填栅和功函数堆叠层,fin之间15-20nm的距离是必要的。“So,youhavethiscliff.”工艺(Foundry)......
  • eval() train() lightning
    https://discuss.pytorch.org/t/why-not-use-model-eval-in-training-step-method-on-lightning/122425/5Issue: IguessPLauthorstookcareofswitchingbetweeneval/trainmodewithinpredefinedcallbacks…ButproblemisthatwhenItrytopredicttestdata......
  • mysql中删除时报错Cannot truncate a table referenced in a foreign key constraint
    在Mysql使用Truncate截断表时,提示Cannottruncateatablereferencedinaforeignkeyconstraint(monitoritem,CONSTRAINTmonitortaskpollutant_monitortask_fk)。这是因为存在外键约束导致的无法删除,我们可以先关闭外键约束,删除后再启动外键约束。1、检查外键约束SELE......
  • 论文阅读笔记《Training Socially Engaging Robots Modeling Backchannel Behaviors w
    TrainingSociallyEngagingRobotsModelingBackchannelBehaviorswithBatchReinforcementLearning训练社交机器人:使用批量强化学习对反馈信号行为进行建模发表于TAC2022。HussainN,ErzinE,SezginTM,etal.TrainingSociallyEngagingRobots:ModelingBackc......
  • 【论文解读】(拼音+字形+字三种信息的中文BERT)ChineseBERT: Chinese Pretraining Enhan
    文章目录1.相关信息2.论文内容3.论文模型3.1GlyphEmbedding3.2PinyinEmbedding4.实验与结论5.模型使用方式1.相关信息论文年份:2021论文地址:https://aclanthology.org/2021.acl-long.161.pdf论文代码(官方):https://github.com/ShannonAI/ChineseBertHuggingFace:Shan......
  • 【论文笔记】MacBert:Revisiting Pre-trained Models for Chinese Natural Language Pr
    文章目录相关信息摘要(Abstract)1.介绍(Introduction)2.相关工作(RelatedWork)3.中文预训练模型(ChinesePre-trainedLanguageModels)3.1BERT-wwm&RoBERTa-wwm3.2MacBERT4.实验设置(ExperimentSetups)4.1SetupsforPre-TrainedLanguageModels4.2SetupsforFine-tuningTask......
  • 论文解读《Interpolated Adversarial Training: Achieving robust neural networks wi
    论文信息论文标题:InterpolatedAdversarialTraining:Achievingrobustneuralnetworkswithoutsacrificingtoomuchaccuracy论文作者:AlexLambVikasVermaKenjiKawaguchiAlexanderMatyaskoSavyaKhoslaJuhoKannalaYoshuaBengio论文来源:2022NeuralNetworks论文地址:dow......
  • 迁移学习(VMT)《Virtual Mixup Training for Unsupervised Domain Adaptation》
    论文信息论文标题:VirtualMixupTrainingforUnsupervisedDomainAdaptation论文作者:TakeruMiyato,S.Maeda,MasanoriKoyama,S.Ishii论文来源:2019CVPR论文地址:download 论文代码:download视屏讲解:click   ......
  • AtCoder Regular Contest 123 E Training
    洛谷传送门AtCoder传送门不妨假设\(B_X\leB_Y\)。设\(f(x)=A_X+\frac{x}{B_X},g(x)=A_Y+\frac{x}{B_Y},F(x)=\left\lfloor{f(x)}\right\rfloor,G(x)=\left\lfloor{g(x)}\right\rfloor\),题目即求\(\sum\limits_{x=1}^n[F(x)=G(x)]\)。我一开始尝试化简......