17 skills found
Nth-iteration-labs / StreamingbanditPython application to setup and run streaming (contextual) bandit experiments.
Nth-iteration-labs / ContextualContextual Bandits in R - simulation and evaluation of Multi-Armed Bandit Policies
dquail / NonStationaryBanditNon stationary bandit for experiments with Reinforcement Learning
Matrix-Factorization-Bandit / IMFB KDD2019Code for the experiments of Matrix Factorization Bandit
v-i-s-h / MAB.jlA Julia Package for providing Multi Armed Bandit Experiments
ir-uam / KNNBanditSoftware for the experiments reported in the RecSys 2019 paper "A Simple Multi-Armed Nearest-Neighbor Bandit for Interactive Recommendation"
aaronsw / PybanditPython experiments in bandit algorithms
ir-uam / EnsembleBanditsSoftware for the experiments reported in the RecSys 2019 paper "Multi-Armed Recommender System Bandit Ensembles"
allenanie / EVOLvE🤖 A framework for experimenting with LLMs in bandit scenarios with customizable agents.
Nth-iteration-labs / Streamingbandit UiClient that handles the administration of StreamingBandit online, or straight from your desktop. Setup and run streaming (contextual) bandit experiments in your browser.
louisfaury / Logistic BanditLogistic Bandit experiments. Official code for the paper "Jointly Efficient and Optimal Algorithms for Logistic Bandits".
djo / Delayed BanditMulti-armed bandit problem under delayed feedback: framework for the numerical experiments
ZhaoHaoRu / Federated Clustering Of BanditsThe experiment for Federated Online Clustering of Bandits (UAI, 2022)
withai / Policy Gradients Contextual Bandit ProblemThe contextual Bandit problem is the intermediate between Simple Bandit problem and the full RL problem. In this experiment we are going to find optimal policy to obtain maximum rewards.
criteo-research / Logistic BanditLogistic Bandit experiments. Official code for the paper "Jointly Efficient and Optimal Algorithms for Logistic Bandits".
alextanhongpin / Go Bandit ServerA bandit server that implements the multi-armed bandit for running a single experiment
paul-english / Multi Armed Bandits ApiA simple API to serve choices and receive feedback for multi-armed bandit type experiments