Published June 27, 2023
| Version v1
Preprint
Open
Fully-Decentralized Training of GNNs using Layer-wise Self-Supervision
Description
In existing literature, GNN training has been performed mostly in centralized, and sometimes federated, settings. In this work, we consider a fully-decentralized data-private scenario, where each node has limited knowledge of the surrounding graph. We propose the first architecture that enables GNN training in this fully-decentralized setting, by carefully combining several techniques, including decoupled learning, self-supervision and Gossip Learning. We implement two simulation tools to experimentally evaluate our solution. The results show that the proposed technique can be effectively used in scenarios where centralized or federated approaches are unfeasible or undesirable.
Files
DecGNN_Paper___V2.pdf
Files
(668.4 kB)
Name | Size | Download all |
---|---|---|
md5:dd7ae002608d7cb48746110929293a2e
|
668.4 kB | Preview Download |