From 08330411ccdbd7de2bd09f6d19c764000ddaaacf Mon Sep 17 00:00:00 2001 From: Vidushi Gupta <55969597+Vidushi-Gupta@users.noreply.github.com> Date: Thu, 8 Jun 2023 15:32:24 +0530 Subject: [PATCH] Updated formula Changed the gif formula to an image --- 8-Reinforcement/1-QLearning/README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/8-Reinforcement/1-QLearning/README.md b/8-Reinforcement/1-QLearning/README.md index ae8a4b5da..38b8e3111 100644 --- a/8-Reinforcement/1-QLearning/README.md +++ b/8-Reinforcement/1-QLearning/README.md @@ -186,7 +186,7 @@ Suppose we are now at the state *s*, and we want to move to the next state *s'*. This gives the **Bellman formula** for calculating the value of the Q-Table at state *s*, given action *a*: - + Here γ is the so-called **discount factor** that determines to which extent you should prefer the current reward over the future reward and vice versa.