| 
 | 1 | +import pytest  | 
 | 2 | + | 
 | 3 | +from rl import *  | 
 | 4 | +from mdp import sequential_decision_environment  | 
 | 5 | + | 
 | 6 | + | 
 | 7 | +north = (0, 1)  | 
 | 8 | +south = (0,-1)  | 
 | 9 | +west = (-1, 0)  | 
 | 10 | +east = (1, 0)  | 
 | 11 | + | 
 | 12 | +policy = {  | 
 | 13 | +    (0, 2): east,  (1, 2): east,  (2, 2): east,   (3, 2): None,  | 
 | 14 | +    (0, 1): north,                (2, 1): north,  (3, 1): None,  | 
 | 15 | +    (0, 0): north, (1, 0): west,  (2, 0): west,   (3, 0): west,   | 
 | 16 | +}  | 
 | 17 | + | 
 | 18 | + | 
 | 19 | + | 
 | 20 | +def test_PassiveADPAgent():  | 
 | 21 | +	agent = PassiveADPAgent(policy, sequential_decision_environment)  | 
 | 22 | +	for i in range(75):  | 
 | 23 | +		run_single_trial(agent,sequential_decision_environment)  | 
 | 24 | +	  | 
 | 25 | +	# Agent does not always produce same results.  | 
 | 26 | +	# Check if results are good enough.  | 
 | 27 | +	assert agent.U[(0, 0)] > 0.15 # In reality around 0.3  | 
 | 28 | +	assert agent.U[(0, 1)] > 0.15 # In reality around 0.4  | 
 | 29 | +	assert agent.U[(1, 0)] > 0 # In reality around 0.2  | 
 | 30 | + | 
 | 31 | + | 
 | 32 | + | 
 | 33 | +def test_PassiveTDAgent():  | 
 | 34 | +	agent = PassiveTDAgent(policy, sequential_decision_environment, alpha=lambda n: 60./(59+n))  | 
 | 35 | +	for i in range(200):  | 
 | 36 | +		run_single_trial(agent,sequential_decision_environment)  | 
 | 37 | +	  | 
 | 38 | +	# Agent does not always produce same results.  | 
 | 39 | +	# Check if results are good enough.  | 
 | 40 | +	assert agent.U[(0, 0)] > 0.15 # In reality around 0.3  | 
 | 41 | +	assert agent.U[(0, 1)] > 0.15 # In reality around 0.35  | 
 | 42 | +	assert agent.U[(1, 0)] > 0.15 # In reality around 0.25  | 
 | 43 | + | 
 | 44 | + | 
 | 45 | +def test_QLearning():  | 
 | 46 | +	q_agent = QLearningAgent(sequential_decision_environment, Ne=5, Rplus=2,   | 
 | 47 | +							 alpha=lambda n: 60./(59+n))  | 
 | 48 | + | 
 | 49 | +	for i in range(200):  | 
 | 50 | +		run_single_trial(q_agent,sequential_decision_environment)  | 
 | 51 | + | 
 | 52 | +	# Agent does not always produce same results.  | 
 | 53 | +	# Check if results are good enough.  | 
 | 54 | +	assert q_agent.Q[((0, 1), (0, 1))] >= -0.5 # In reality around 0.1  | 
 | 55 | +	assert q_agent.Q[((1, 0), (0, -1))] <= 0.5 # In reality around -0.1  | 
0 commit comments