SkillAgentSearch skills...

Pconf

[NeurIPS 2018] Code for the paper "Binary Classification from Positive-Confidence Data"

Install / Use

/learn @takashiishida/Pconf
About this skill

Quality Score

0/100

Supported Platforms

Universal

README

Positive-Confidence (Pconf) Classification

NeurIPS 2018 Poster Slides MIT License

This repository gives the implementation for Positive-Confidence (Pconf) classification that was proposed in the paper "Binary Classification from Positive-Confidence Data" (NeurIPS 2018).

Implementation

We use two Gaussian distributions for the positive and negative classes. For Guassian distributions, we can analytically calculate the positive posterior which we use for the confidence values. We perform Pconf classification and compare it with several baselines: Naive weighted classification, one-class SVM, and fully-supervised classification. These classifiers are implemented in components.py.

This code can be used to reproduce the synthetic experiments with true positive-confidence in Section 4.1 of our paper. Note that the synthetic experiments with noisy positive-confidence can be conducted by adding zero-mean Gaussian noise to the true confidence.

Requirements

  • Python 3.6
  • numpy 1.13
  • Pytorch 0.4
  • matplotlib 2.1.1

Getting started

$ python visualize.py

Expected results

References

For the details, please see the paper here.

@article{pconf,
	author = {Takashi Ishida and Gang Niu and Masashi Sugiyama},
	title = {Binary Classification from Positive-Confidence Data},
	booktitle = {NeurIPS},
	year = {2018}
}
View on GitHub
GitHub Stars51
CategoryEducation
Updated6mo ago
Forks7

Languages

Python

Security Score

92/100

Audited on Sep 5, 2025

No findings