Multi-armed Bandits with Constraint

R∗ is the expected reward of the optimal policy (the policy that can obtain the maximum expected reward given the reward and cost distributions of each arm),
  1. For reward- Beta(1,1)
  2. For cost- Beta(1,1)
arms with cost -10,8,5,3, iteration 200
iteration 3000
import numpy as np
import random
from scipy.stats import beta
import matplotlib.pyplot as plt
from numpy.random import choice
plt.rcParams['figure.figsize'] = [10, 7]
import pandas as pd
def plot_list(reward,penalties,iteration,plot_name):
x = np.linspace(0, 1, 10000)
y1 = beta.pdf(x, reward[0], penalties[0])
y2 = beta.pdf(x, reward[1], penalties[1])
y3 = beta.pdf(x, reward[2], penalties[2])
y4 = beta.pdf(x, reward[3], penalties[3])
plt.title(f"PDF of Beta (Bell-shape) iteration({iteration} // {plot_name})", fontsize=20)
plt.xlabel("X", fontsize=16)
plt.ylabel("Probability Density", fontsize=16)
plt.plot(x, y1, linewidth=3, color='firebrick')
plt.annotate(f"Beta( {reward[0]},{penalties[0]}) { 10}", xy=(1.75, 19), size = 14, ha='center', va='center', color='firebrick')
plt.plot(x, y2, linewidth=3, color='burlywood')
plt.annotate(f"Beta( {reward[1]},{penalties[1]}) { 8}", xy=(1.75, 16.25), size = 14, ha='center', va='center', color='burlywood')
plt.plot(x, y3, linewidth=3, color='dodgerblue')
plt.annotate(f"Beta( {reward[2]},{penalties[2]}) { 5}", xy=(1.75, 15), size = 14, ha='center', va='center', color='dodgerblue')
plt.plot(x, y4, linewidth=3, color='green')
plt.annotate(f"Beta( {reward[3]},{penalties[3]}) { 3}", xy=(1.75, 17.5), size = 14, ha='center', va='center', color='green')
plt.ylim([0, 20])
plt.xlim([0, 2])
def plot_count(num,iteration):['A10','A08','A05','A03'],num)
plt.title(f"Arms selected the most {iteration}")
plt.ylabel('Number Of Times Each Arm Was Selected To Play')
rl10 = [choice([0,0,1]) for i in range(4000)]
rl8 = [choice([0,0,1]) for i in range(4000)]
rl5 = [choice([0,1]) for i in range(4000)]
rl3 = [choice([0,0,1]) for i in range(4000)]
rl10 = [choice([j/10 for j in range(10)]) for i in range(4000)]
rl8 = [choice([j/10 for j in range(10)]) for i in range(4000)]
rl5 = [choice([j/10 for j in range(10)]) for i in range(4000)]
rl3 = [choice([j/10 for j in range(10)]) for i in range(4000)]
def paper_approach_plot(observations,arms,data,data_cost):
#other initializations

arms_selected = []

for n in range(observations):
arm = 0
max_val = 0
for i in range(arms):
if I>max_val:

r = data.values[n, arm]
c = data_2.values[n, arm]
# if r==0:
# c=0

if n%200==0 or n==observations-1:




Connecting the Dots.

Love podcasts or audiobooks? Learn on the go with our new app.

Recommended from Medium

Udacity Deep Learning Nanodegree Notes and Thoughts [Lesson 3]

What to Watch Next? A Basic Recommender System Using Ideas from NLP.

Processing data for Machine Learning with TensorFlow

look into codes

[ML UTD 20] Machine Learning Up-To-Date

TensorFlow vs PyTorch for Deep Learning

Natural language processing is the future assistive educational technology for the intellectually…

Article 1: Understanding the Convolution function and CNN.

Weekly Supervised Learning — Getting started with unstructured data

Get the Medium app

A button that says 'Download on the App Store', and if clicked it will lead you to the iOS App store
A button that says 'Get it on, Google Play', and if clicked it will lead you to the Google Play store
Aashay Sachdeva

Aashay Sachdeva

Connecting the Dots.

More from Medium

Reinforcement Learning with Multi-Armed Bandits

Recommender System Using Reinforcement Learning

Data-centric AI: Practical implications with the SMART Pipeline

Model warmup