From 08330411ccdbd7de2bd09f6d19c764000ddaaacf Mon Sep 17 00:00:00 2001
From: Vidushi Gupta <55969597+Vidushi-Gupta@users.noreply.github.com>
Date: Thu, 8 Jun 2023 15:32:24 +0530
Subject: [PATCH] Updated formula
Changed the gif formula to an image
---
8-Reinforcement/1-QLearning/README.md | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
diff --git a/8-Reinforcement/1-QLearning/README.md b/8-Reinforcement/1-QLearning/README.md
index ae8a4b5da..38b8e3111 100644
--- a/8-Reinforcement/1-QLearning/README.md
+++ b/8-Reinforcement/1-QLearning/README.md
@@ -186,7 +186,7 @@ Suppose we are now at the state *s*, and we want to move to the next state *s'*.
This gives the **Bellman formula** for calculating the value of the Q-Table at state *s*, given action *a*:
-
+
Here γ is the so-called **discount factor** that determines to which extent you should prefer the current reward over the future reward and vice versa.