关于q-learning算法,可参照以下博客,我只是复现作者的算法,如有错误,请私信改正。
A Painless Q-learning Tutorial (一个 Q-learning 算法的简明教程)_peghoty-CSDN博客
import numpy as np
import pandas as pd
class QL:
def __init__(self, actions, gamma=0.8, e_greedy=0.9):
self.actions = actions
self.gamma = gamma
self.e_greedy = e_greedy
self.q_table = pd.DataFrame(columns=actions, dtype=np.float64) #行为状态,列是动作,这里简化了,同时也是下一个状态
def choose_action(self, state):
self.check_state(state)
if(np.random.uniform(0, 1)<self.e_greedy):
action_list = self.q_table.loc[state, :] # 取出当前state的行
action = action_list[action_list == action_list.max()].index #找最大值的动作,可能有多个,比如刚开始都是0.
action = np.random.choice(action) # 在最大值中随机选择一个动作。
else:
action = np.random.choice(self.actions)
return action
def learn(self, state_now, state_next, reward_value):
state_next_list = self.q_table.loc[state_next, :]
self.q_table.loc[state_now, state_next] = reward_value + self.gamma * state_next_list.max()
def check_state(self, state):
if state not in self.q_table.index:
self.q_table = self.q_table.append(pd.Series([0]*len(self.actions), index=self.actions,name=state))
terminal = 5 # 出口在5
times = 1000 # 进行1000次的尝试
actions = np.array([0,1,2,3,4,5]) #6个动作
agent = QL(actions) # 6个动作
reward_rect = [[-1,-1,-1,-1,0,-1],
[-1,-1,-1,0,-1,100],
[-1,-1,-1,0,-1,-1],
[-1,0,0,-1,0,-1],
[0,-1,-1,0,-1,100],
[-1,0,-1,-1,0,100]]
reward_rect = np.array(reward_rect) # 收益矩阵R
for episode in range(times):
state_now = np.random.choice(agent.actions) # 每次开始都随机选择一个初始状态作为起点
agent.check_state(state_now) # 检查有没有到过这个状态
route = [state_now] # 该列表记录到达过的状态,运行的时候,可以先把迭代次数调至50,把route可以打印出来看看(在下面取消注释就行了)
# 开始探索
while True:
state_next = agent.choose_action(state_now) # 寻找下一状态
agent.check_state(state_next)
route.append(state_next)
if(reward_rect[state_now, state_next]==-1):
break
agent.learn(state_now, state_next, reward_rect[state_now, state_next])
if(state_next==terminal):
break
state_now = state_next
# print("行走路线:", route)
print(agent.q_table)