SkillAgentSearch skills...

GraphSmote

Pytorch implementation of paper 'GraphSMOTE: Imbalanced Node Classification on Graphs with Graph Neural Networks' to appear on WSDM2021

Install / Use

/learn @TianxiangZhao/GraphSmote
About this skill

Quality Score

0/100

Supported Platforms

Universal

README

GraphSmote

Pytorch implementation of paper 'GraphSMOTE: Imbalanced Node Classification on Graphs with Graph Neural Networks' on WSDM2021

Dependencies

CPU

  • python3
  • ipdb
  • pytorch1.0
  • network 2.4
  • scipy
  • sklearn

Dataset

Two processed datasets are published, including BlogCatalog and Cora. For downloading them, please click. Please unzip them in the 'data' folder.

The twitter dataset is obtained from link, with fake id. For twitter dataset, node embeddings can be generated with Deepwalk, following this repo. Loading the twitter dataset is also provided in data_load.py.

Configurations

Architectures

We provide two base architectures, GCN and GraphSage. The default one is GraphSage, and can be set via '--model'.

Upscale ratios

The default value is 1. If want to make every class balanced instead of using pre-set ratios, please set it to 0 in '--up_scale'.

Finetuning the decoder

During finetune, set '--setting='newG_cls'' correponds to use pretrained decoder, and set '--setting='recon_newG'' corresponds to also finetune the decoder.

Besides, during finetune, '--opt_new_G' corresponds to update decoder with also classification losses. This option may cause more variance in performance, and usually need more careful hyper-parameter choices.

GraphSMOTE

Below is an example for the Cora dataset.

Train

  • Pretrain the auto-encoder

<code>python main.py --imbalance --no-cuda --dataset=cora --setting='recon'</code>

Pretrained model can be found in the corresponding checkpoint folder. Rename and set the path to pretrained checkpoint as [dataset]\Pretrained.pth

  • Finetune

<code>python main.py --imbalance --no-cuda --dataset=cora --setting='newG_cls' --load=Pretrained.pth</code>

Baselines

We provide four baselines in this code. They can be configured via the '--setting' arguments. Please refer to the 'get_parser()' function in utils.py.

  • Oringinal model: Vanilla backbone models. '--setting='no''
  • Over-sampling in raw input domain: Repeat nodes in the minority classes. '--setting='upsampling''
  • Reweight: Give samples from minority classes a larger weight when calculating the loss. '--setting='reweight''
  • Embed-SMOTE: Perform SMOTE in the intermediate embedding domain. '--setting='embed_up''

Use over-sampling as an example:

<code>python main.py --imbalance --no-cuda --dataset=cora --setting='upsampling'</code>

Citation

If any problems occur via running this code, please contact us at tkz5084@psu.edu.

Thank you!

Related Skills

View on GitHub
GitHub Stars176
CategoryDevelopment
Updated3d ago
Forks35

Languages

Python

Security Score

80/100

Audited on Mar 23, 2026

No findings