Need help with CorrectAndSmooth?
Click the “chat” button below for chat support from the developer who created it, or find similar developers for support.

About the developer

207 Stars 38 Forks 17 Commits 5 Opened issues


[ICLR 2021] Combining Label Propagation and Simple Models Out-performs Graph Neural Networks (

Services available


Need anything else?

Contributors list

Correct and Smooth (C&S) OGB submissions


This directory contains OGB submissions. All hyperparameters were tuned on the validation set with optuna, except for products, which was hand tuned. All experiments were run with a RTX 2080 TI with 11GB.

Some Tips

  • In general, the more complex and "smooth" your GNN is, the less likely it'll be that applying the "Correct" portion helps performance. In those cases, you may consider just applying the "smooth" portion, like we do on the GAT. In almost all cases, applying the "smoothing" component will improve performance. For Linear/MLP models, applying the "Correct" portion is almost always essential for obtaining good performance.

  • In a similar vein, an improvement of performance of your model may not correspond to an improvement after applying C&S. Considering that C&S learns no parameters over your data, our intuition is that C&S "levels" the playing field, allowing models that learn interesting features to shine (as opposed to learning how to be smooth).

    • Even though GAT (73.57) is outperformed by GAT + labels (73.65), when we apply C&S, we see that GAT + C&S (73.86) performs better than GAT + labels + C&S (~73.70) ,
    • Even though a 6 layer GCN performs on par with a 2 layer GCN with Node2Vec features, C&S improves performance of the 2 layer GCN with Node2Vec features substantially more.
    • Even though MLP + Node2Vec outperforms MLP + Spectral in both arxiv and products, the performance ordering flips after we apply C&S.
    • On Products, the MLP (74%) is substantially outperformed by ClusterGCN (80%). However, MLP + C&S (84.1%) substantially outperforms ClusterGCN + C&S (82.4%).
  • In general, autoscale works more reliably than fixedscale, even though fixedscale may make more sense...


Label Propagation (0 params):

python --dataset arxiv --method lp

Valid acc: 0.7013658176448874 Test acc: 0.6832294302820814

Plain Linear + C&S (5160 params, 52.5% base accuracy)

python --dataset arxiv --model plain --epochs 1000    
python --dataset arxiv --method plain

Valid acc -> Test acc Args []: 73.00 ± 0.01 -> 71.26 ± 0.01

Linear + C&S (15400 params, 70.11% base accuracy)

python --dataset arxiv --model linear --use_embeddings --epochs 1000 
python --dataset arxiv --method linear

Valid acc -> Test acc Args []: 73.68 ± 0.04 -> 72.22 ± 0.02;

MLP + C&S (175656 params, 71.44% base accuracy)

python --dataset arxiv --model mlp --use_embeddings
python --dataset arxiv --method mlp

Valid acc -> Test acc Args []: 73.91 ± 0.15 -> 73.12 ± 0.12

GAT + C&S (1567000 params, 73.56% base accuracy)

cd gat && python --use-norm
cd .. && python --dataset arxiv --method gat

Valid acc -> Test acc Args []: 74.84 ± 0.07 -> 73.86 ± 0.14


As opposed to the paper's results, which only use spectral embeddings, here we use spectral and diffusion embeddings, which we find improves Arxiv performance.


Label Propagation (0 params):

python --dataset products --method lp 

Valid acc: 0.9090608549703736 Test acc: 0.7434145274640762

Plain Linear + C&S (4747 params, 47.73% base accuracy)

python --dataset products --model plain --epochs 1000 --lr 0.1
python --dataset products --method plain

Valid acc -> Test acc Args []: 91.03 ± 0.01 -> 82.54 ± 0.03

Linear + C&S (10763 params, 50.05% base accuracy)

python --dataset products --model linear --use_embeddings --epochs 1000 --lr 0.1
python --dataset products --method linear

Valid acc -> Test acc Args []: 91.34 ± 0.01 -> 83.01 ± 0.01

MLP + C&S (96247 params, 63.41% base accuracy)

python --dataset products --model mlp --hidden_channels 200 --use_embeddings
python --dataset products --method mlp

Valid acc -> Test acc Args []: 91.47 ± 0.09 -> 84.18 ± 0.07

We use cookies. If you continue to browse the site, you agree to the use of cookies. For more information on our use of cookies please see our Privacy Policy.