This repo is for source code of WWW 2022 paper "Confidence May Cheat: Self-Training on Graph Neural Networks under Distribution Shift".
Paper Link: https://arxiv.org/abs/2201.11349
- python == 3.8.8
- pytorch == 1.8.1
- dgl -cuda11.1 == 0.6.1
- networkx == 2.5
- numpy == 1.20.2
python main.py --dataset dataset --m model --labelrate labelrate --s stage --t threshold --b beta --drop_method drop_method --droprate droprate
- dataset: including [Cora, Citeseer, Pubmed, CoraFull, Flickr]
- model: including [GCN, GAT, APPNP]
- labelrate: including [3, 5, 10, 20]
- stage: ranging from 1 to 10
- drop_method: including [dropout, dropedge] e.g.
python main.py --dataset Citeseer --m GCN --labelrate 20 --s 9 --t 0.7 --b 0.5 --drop_method dropout --droprate 0.3
Dataset-labelrate-threshold-beta-droprate | |
---|---|
Cora-20-0.530000-0.333333-0.500000-dropout | stage=2 |
Citeseer-20-0.700000-0.500000-0.300000-dropout | stage=9 |
Pubmed-20-0.740000-1.000000-0.500000-dropout | stage=2 |
CoraFull-20-0.800000-0.250000-0.400000-dropout | stage=5 |
Flickr-20-0.890000-1.000000-0.100000-dropout | stage=9 |
Dataset-labelrate-threshold-beta-droprate | |
---|---|
Cora-20-0.540000-0.500000-0.400000-dropedge | stage = 5 |
Citeseer-20-0.600000-0.500000-0.100000-dropedge | stage = 3 |
Pubmed-20-0.710000-1.000000-0.500000-dropedge | stage = 2 |
CoraFull-20-0.910000-0.200000-0.500000-dropedge | stage = 9 |
Flickr-20-0.950000-0.500000-0.200000-dropedge | stage = 6 |
Dataset-labelrate-threshold-beta-droprate | |
---|---|
Cora-10-0.700000-0.333333-0.300000-dropout | stage=6 |
Citeseer-10-0.550000-0.500000-0.300000-dropout | stage=4 |
Pubmed-10-0.750000-1.000000-0.300000-dropout | stage=1 |
CoraFull-10-0.780000-0.333333-0.100000-dropout | stage=6 |
Flickr-10-0.960000-1.000000-0.100000-dropout | stage=7 |
Dataset-labelrate-threshold-beta-droprate | |
---|---|
Cora-10-0.650000-0.500000-0.500000-dropedge | stage = 3 |
Citeseer-10-0.650000-0.500000-0.300000-dropedge | stage = 6 |
Pubmed-10-0.600000-0.666667-0.300000-dropedge | stage = 1 |
CoraFull-10-0.630000-0.333333-0.300000-dropedge | stage = 2 |
Flickr-10-0.930000-0.666667-0.100000-dropedge | stage = 2 |
Dataset-labelrate-threshold-beta-droprate | |
---|---|
Cora-5-0.600000-0.500000-0.500000-dropout | stage=2 |
Citeseer-5-0.450000-0.666667-0.500000-dropout | stage=9 |
Pubmed-5-0.870000-1.000000-0.300000-dropout | stage=1 |
CoraFull-5-0.690000-0.333333-0.100000-dropout | stage=2 |
Flickr-5-0.930000-1.000000-0.100000-dropout | stage=2 |
Dataset-labelrate-threshold-beta-droprate | |
---|---|
Cora-5-0.600000-0.500000-0.300000-dropedge | stage = 2 |
Citeseer-5-0.400000-0.500000-0.100000-dropedge | stage = 3 |
Pubmed-5-0.600000-1.000000-0.300000-dropedge | stage = 1 |
CoraFull-5-0.600000-0.500000-0.300000-dropedge | stage = 2 |
Flickr-5-0.960000-1.000000-0.500000-dropedge | stage = 4 |
Dataset-labelrate-threshold-beta-droprate | |
---|---|
Cora-3-0.700000-0.500000-0.300000-dropout | stage=3 |
Citeseer-3-0.450000-0.666667-0.300000-dropout | stage=2 |
Pubmed-3-0.780000-0.500000-0.500000-dropout | stage=1 |
CoraFull-3-0.870000-0.333333-0.300000-dropout | stage=2 |
Flickr-3-0.960000-1.000000-0.500000-dropout | stage=4 |
Dataset-labelrate-threshold-beta-droprate | |
---|---|
Cora-3-0.400000-0.333333-0.100000-dropedge | stage = 2 |
Citeseer-3-0.450000-0.333333-0.500000-dropedge | stage = 1 |
Pubmed-3-0.810000-0.500000-0.300000-dropedge | stage = 1 |
CoraFull-3-0.660000-0.500000-0.300000-dropedge | stage = 1 |
Flickr-3-0.900000-0.666667-0.500000-dropedge | stage = 1 |
If you have any questions, please feel free to contact me with [email protected]