This commit is contained in:
Zheyuan Wu
2025-10-14 20:34:47 -05:00
parent 250f763f1f
commit e74aac95e3
517 changed files with 1418 additions and 16701 deletions

View File

@@ -2,7 +2,7 @@ import os
import torch
import torch.optim as optim
from copy import deepcopy
from model import QNetwork, DuelingQNetwork
from model import QNetwork, DuelingQNetwork, NoisyQNetwork
from gymnasium.wrappers import TimeLimit
class DQNAgent:
@@ -10,10 +10,17 @@ class DQNAgent:
self.device = device
self.use_double = cfg.use_double
self.use_dueling = cfg.use_dueling
self.use_noisy = cfg.use_noisy
self.noisy_sigma = cfg.noisy_sigma
self.target_update_interval = cfg.target_update_interval
q_model = DuelingQNetwork if self.use_dueling else QNetwork
self.q_net = q_model(state_size, action_size, cfg.hidden_size, cfg.activation).to(self.device)
q_model = QNetwork
if self.use_dueling:
q_model = DuelingQNetwork
if self.use_noisy:
q_model = NoisyQNetwork
self.q_net = q_model(state_size, action_size, cfg.hidden_size, cfg.activation, sigma_init=cfg.noisy_sigma).to(self.device)
else:
self.q_net = q_model(state_size, action_size, cfg.hidden_size, cfg.activation).to(self.device)
self.target_net = deepcopy(self.q_net).to(self.device)
self.optimizer = optim.AdamW(self.q_net.parameters(), lr=cfg.lr)
@@ -51,12 +58,14 @@ class DQNAgent:
if self.use_double:
# YOUR IMPLEMENTATION HERE
reward_tensor = reward.to(self.device)
# update from batch states via q_net
next_q_tensor = self.q_net(next_state.to(self.device))
next_q_tensor = self.target_net(next_state.to(self.device))
next_action = torch.argmax(self.q_net(next_state.to(self.device)), dim=1).unsqueeze(1)
# print(next_q_tensor.shape, next_action.shape)
# return the max Q value
next_q = torch.max(next_q_tensor, dim=1).values
next_q = torch.gather(next_q_tensor, dim=1, index=next_action).squeeze(1)
q_target = reward_tensor + (1-done.to(self.device)) * self.gamma * next_q
return q_target
else:
# YOUR IMPLEMENTATION HERE
reward_tensor = reward.to(self.device)
@@ -73,22 +82,14 @@ class DQNAgent:
"""
############################
# YOUR IMPLEMENTATION HERE #
if use_double_net:
# get from target net
q_tensor = self.target_net(state.to(self.device))
action_idx = action.squeeze(1).to(dtype=torch.int32).to(self.device)
# select corresponding action, do not use index_select... That don't works
q = q_tensor.gather(1, action_idx.unsqueeze(1)).squeeze(1)
return q
else:
# elegant python move by Jack Wu. Fantastic...
# q= self.q_net(state.to(self.device))[:, action.int()]
# update from batch states
q_tensor = self.q_net(state.to(self.device))
action_idx = action.squeeze(1).to(dtype=torch.int32).to(self.device)
# select corresponding action, do not use index_select... That don't works
q = q_tensor.gather(1, action_idx.unsqueeze(1)).squeeze(1)
return q
# elegant python move by Jack Wu. Fantastic...
# q= self.q_net(state.to(self.device))[:, action.int()]
# update from batch states
q_tensor = self.q_net(state.to(self.device))
action_idx = action.squeeze(1).to(dtype=torch.int32).to(self.device)
# select corresponding action, do not use index_select... That don't works
q = q_tensor.gather(1, action_idx.unsqueeze(1)).squeeze(1)
return q
############################
def update(self, batch, step, weights=None):
@@ -123,5 +124,7 @@ class DQNAgent:
def __repr__(self) -> str:
use_double = 'Double' if self.use_double else ''
use_dueling = 'Dueling' if self.use_dueling else ''
prefix = 'Normal' if not self.use_double and not self.use_dueling else ''
return use_double + use_dueling + prefix + 'QNetwork'
use_noisy = 'Noisy' if self.use_noisy else ''
prefix = 'Normal' if not self.use_double and not self.use_dueling and not self.use_noisy else ''
suffix = f'with noisy sigma={self.noisy_sigma}' if self.use_noisy else ''
return use_double + use_dueling + use_noisy+ prefix + 'QNetwork' + suffix

View File

@@ -85,14 +85,13 @@ class NStepReplayBuffer(ReplayBuffer):
"""Get n-step state, action, reward and done for the transition, discard those rewards after done=True"""
############################
# YOUR IMPLEMENTATION HERE #
state, action, reward, done = self.n_step_buffer[0]
state, action, reward, done = self.n_step_buffer.popleft()
# compute n-step discounted reward
gamma = self.gamma
for i in range(1, len(self.n_step_buffer)):
if done:
for i in range(self.n_step - 1):
reward += self.gamma**(i+1) * self.n_step_buffer[i][2]
# ignore done steps
if self.n_step_buffer[i][3]:
break
reward += gamma * self.n_step_buffer[i][2]
gamma *= self.gamma
############################
return state, action, reward, done
@@ -192,11 +191,12 @@ class PrioritizedNStepReplayBuffer(PrioritizedReplayBuffer):
# YOUR IMPLEMENTATION HERE #
state, action, reward, done = self.n_step_buffer[0]
# compute n-step discounted reward
gamma = self.gamma
for i in range(1, len(self.n_step_buffer)):
if done:
state, action, reward, done = self.n_step_buffer.popleft()
# compute n-step discounted reward
for i in range(self.n_step - 1):
reward += self.gamma**(i+1) * self.n_step_buffer[i][2]
# ignore done steps
if self.n_step_buffer[i][3]:
break
reward += gamma * self.n_step_buffer[i][2]
gamma *= self.gamma
############################
return state, action, reward, done

View File

@@ -26,6 +26,8 @@ agent:
# you can define other parameters of the __init__ function (if any) for the object here
use_dueling: False
use_double: False
use_noisy: False
noisy_sigma: 0.5
buffer:
capacity: 50_000

1
hw2/commands/4-8.sh Normal file
View File

@@ -0,0 +1 @@
python main.py agent.use_noisy=true agent.noisy_sigma=0.017

View File

@@ -2,6 +2,10 @@ from hydra.utils import instantiate
import torch
import torch.nn as nn
# additional imports for extra credit
import math
import torch.nn.functional as F
class QNetwork(nn.Module):
def __init__(self, state_size, action_size, hidden_size, activation):
@@ -49,5 +53,55 @@ class DuelingQNetwork(nn.Module):
############################
return Qs
# Extra credit: implementing Noisy DQN
class NoisyLinear(nn.Linear):
# code reference from:
# (1) https://github.com/PacktPublishing/Deep-Reinforcement-Learning-Hands-On/blob/baa9d013596ea8ea8ed6826b9de6679d98b897ca/Chapter07/lib/dqn_model.py#L9
# (2) https://github.com/thomashirtz/noisy-networks/blob/main/noisynetworks.py
def __init__(self, in_features, out_features, sigma_init=0.5, bias=True):
super().__init__(in_features, out_features, bias=bias)
# assume noise is gaussian, set sigma as learnable parameters
self.sigma_weight = nn.Parameter(torch.full((out_features, in_features), sigma_init))
self.register_buffer('epsilon_weight', torch.full((out_features, in_features), sigma_init))
if bias:
self.sigma_bias = nn.Parameter(torch.full((out_features,), sigma_init))
self.register_buffer('epsilon_bias', torch.full((out_features,), sigma_init))
self.reset_parameters()
def reset_parameters(self):
"""
Reset the weights and bias of the noisy linear layer to a uniform distribution with std dev of sqrt(3 / in_features)
"""
std = math.sqrt(3 / self.in_features)
self.weight.data.uniform_(-std, std)
self.bias.data.uniform_(-std, std)
def forward(self, input):
"""
Forward pass of noisy linear layer, adding gaussian noise to the weight and bias
"""
self.epsilon_weight.normal_()
weight = self.weight + self.sigma_weight * self.epsilon_weight.data
bias = self.bias
if bias is not None:
self.epsilon_bias.normal_()
bias = bias + self.sigma_bias * self.epsilon_bias.data
return F.linear(input, weight, bias)
class NoisyQNetwork(nn.Module):
def __init__(self, state_size, action_size, hidden_size, activation, sigma_init=0.5):
super(NoisyQNetwork, self).__init__()
self.q_head = nn.Sequential(
NoisyLinear(state_size, hidden_size, sigma_init=sigma_init),
instantiate(activation),
NoisyLinear(hidden_size, hidden_size, sigma_init=sigma_init),
instantiate(activation),
NoisyLinear(hidden_size, action_size, sigma_init=sigma_init)
)
def forward(self, state):
Qs = self.q_head(state)
return Qs

View File

@@ -1,5 +1,15 @@
\relax
\newlabel{1}{{{1}}{1}{}{}{}}
\newlabel{2}{{{2}}{1}{}{}{}}
\newlabel{3}{{{3}}{1}{}{}{}}
\gdef \@abspage@last{4}
\providecommand\hyper@newdestlabel[2]{}
\providecommand\HyField@AuxAddToFields[1]{}
\providecommand\HyField@AuxAddToCoFields[2]{}
\newlabel{1}{{{1}}{2}{}{AMS.1}{}}
\newlabel{2}{{{2}}{2}{}{AMS.2}{}}
\newlabel{3}{{{3}}{2}{}{AMS.3}{}}
\@writefile{lof}{\contentsline {figure}{\numberline {1}{\ignorespaces DQN. Nothing to say but what expected from training.}}{4}{figure.1}\protected@file@percent }
\@writefile{lof}{\contentsline {figure}{\numberline {2}{\ignorespaces Double DQN. I found there is interesting camel like bump for q-value when training with Double DQN. It is less stable than the vanilla DQN.}}{4}{figure.2}\protected@file@percent }
\@writefile{lof}{\contentsline {figure}{\numberline {3}{\ignorespaces Dueling DQN. Using Advantage network creates comparable results as the DQN.}}{4}{figure.3}\protected@file@percent }
\@writefile{lof}{\contentsline {figure}{\numberline {4}{\ignorespaces Prioritized Experience Replay. Using this alone makes the training process less stable and loss is significantly higher than the previous methods.}}{5}{figure.4}\protected@file@percent }
\@writefile{lof}{\contentsline {figure}{\numberline {5}{\ignorespaces N-Step Experience Replay. So far the most stable method of training, especially when the replay buffer size is large. However, when the replay buffer size is too small, typically $\le 70$, the training process may not converge to optimal performance.}}{5}{figure.5}\protected@file@percent }
\@writefile{lof}{\contentsline {figure}{\numberline {6}{\ignorespaces NStep + PER. Combining the two methods counter the unstable loss function for training in PER.}}{6}{figure.6}\protected@file@percent }
\@writefile{lof}{\contentsline {figure}{\numberline {7}{\ignorespaces Noisy DQN. Experiment for sigma = 0.017 gets comparable result with normal DQN. Stability issue persist when sigma is too large.}}{6}{figure.7}\protected@file@percent }
\gdef \@abspage@last{7}

View File

@@ -1,6 +1,8 @@
# Fdb version 4
["pdflatex"] 1760230164.51698 "d:/Documents/Nextcloud/Documents/Project WUSTL/Academic/2025_Fall/CSE5100/Homeworks/hw2/result.tex" "result.pdf" "result" 1760230165.61579 0
["pdflatex"] 1760491970.16853 "d:/Documents/Nextcloud/Documents/Project WUSTL/Academic/2025_Fall/CSE5100/Homeworks/hw2/result.tex" "result.pdf" "result" 1760491972.62655 0
"c:/texlive/2023/texmf-dist/fonts/enc/dvips/cm-super/cm-super-ts1.enc" 1708989547 2900 1537cc8184ad1792082cd229ecc269f4 ""
"c:/texlive/2023/texmf-dist/fonts/map/fontname/texfonts.map" 1708990624 3524 cb3e574dea2d1052e39280babc910dc8 ""
"c:/texlive/2023/texmf-dist/fonts/tfm/jknappen/ec/tcrm1095.tfm" 1708990172 1536 02c06700a42be0f5a28664c7273f82e7 ""
"c:/texlive/2023/texmf-dist/fonts/tfm/public/amsfonts/cmextra/cmex7.tfm" 1708988591 1004 54797486969f23fa377b128694d548df ""
"c:/texlive/2023/texmf-dist/fonts/tfm/public/amsfonts/cmextra/cmex8.tfm" 1708988591 988 bdf658c3bfc2d96d3c8b02cfc1c94c20 ""
"c:/texlive/2023/texmf-dist/fonts/tfm/public/amsfonts/symbols/msam10.tfm" 1708988591 916 f87d7c45f9c908e672703b83b72241a3 ""
@@ -19,6 +21,7 @@
"c:/texlive/2023/texmf-dist/fonts/tfm/public/cm/cmsy6.tfm" 1708989536 1116 933a60c408fc0a863a92debe84b2d294 ""
"c:/texlive/2023/texmf-dist/fonts/tfm/public/cm/cmsy8.tfm" 1708989536 1120 8b7d695260f3cff42e636090a8002094 ""
"c:/texlive/2023/texmf-dist/fonts/tfm/public/cm/cmti10.tfm" 1708989536 1480 aa8e34af0eb6a2941b776984cf1dfdc4 ""
"c:/texlive/2023/texmf-dist/fonts/tfm/public/cm/cmtt10.tfm" 1708989536 768 1321e9409b4137d6fb428ac9dc956269 ""
"c:/texlive/2023/texmf-dist/fonts/tfm/public/rsfs/rsfs10.tfm" 1708993366 688 37338d6ab346c2f1466b29e195316aa4 ""
"c:/texlive/2023/texmf-dist/fonts/tfm/public/rsfs/rsfs5.tfm" 1708993366 684 3a51bd4fd9600428d5264cf25f04bb9a ""
"c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmbx10.pfb" 1708988591 34811 78b52f49e893bcba91bd7581cdc144c0 ""
@@ -31,9 +34,22 @@
"c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmsy6.pfb" 1708988591 32587 65067f817f408bc71a7312f3d9828a9b ""
"c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmsy8.pfb" 1708988591 32626 5abc8bb2f28aa647d4c70f8ea38cc0d3 ""
"c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmti10.pfb" 1708988591 37944 359e864bd06cde3b1cf57bb20757fb06 ""
"c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmtt10.pfb" 1708988591 31099 342ef5a582aacbd3346f3cf4579679fa ""
"c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/symbols/msbm10.pfb" 1708988591 34694 870c211f62cb72718a00e353f14f254d ""
"c:/texlive/2023/texmf-dist/fonts/type1/public/cm-super/sfrm1095.pfb" 1708989547 145929 5c9aebea9ba6e33fc93158c04a3bdcd8 ""
"c:/texlive/2023/texmf-dist/tex/context/base/mkii/supp-pdf.mkii" 1708992232 71627 94eb9990bed73c364d7f53f960cc8c5b ""
"c:/texlive/2023/texmf-dist/tex/generic/atbegshi/atbegshi.sty" 1708988730 24708 5584a51a7101caf7e6bbf1fc27d8f7b1 ""
"c:/texlive/2023/texmf-dist/tex/generic/bigintcalc/bigintcalc.sty" 1708989153 40635 c40361e206be584d448876bba8a64a3b ""
"c:/texlive/2023/texmf-dist/tex/generic/bitset/bitset.sty" 1708989167 33961 6b5c75130e435b2bfdb9f480a09a39f9 ""
"c:/texlive/2023/texmf-dist/tex/generic/gettitlestring/gettitlestring.sty" 1708990766 8371 9d55b8bd010bc717624922fb3477d92e ""
"c:/texlive/2023/texmf-dist/tex/generic/iftex/iftex.sty" 1708991184 7237 bdd120a32c8fdb4b433cf9ca2e7cd98a ""
"c:/texlive/2023/texmf-dist/tex/generic/infwarerr/infwarerr.sty" 1708991215 8356 7bbb2c2373aa810be568c29e333da8ed ""
"c:/texlive/2023/texmf-dist/tex/generic/intcalc/intcalc.sty" 1708991235 31769 002a487f55041f8e805cfbf6385ffd97 ""
"c:/texlive/2023/texmf-dist/tex/generic/kvdefinekeys/kvdefinekeys.sty" 1708991455 5412 d5a2436094cd7be85769db90f29250a6 ""
"c:/texlive/2023/texmf-dist/tex/generic/ltxcmds/ltxcmds.sty" 1708991801 17865 1a9bd36b4f98178fa551aca822290953 ""
"c:/texlive/2023/texmf-dist/tex/generic/pdfescape/pdfescape.sty" 1708992674 19007 15924f7228aca6c6d184b115f4baa231 ""
"c:/texlive/2023/texmf-dist/tex/generic/pdftexcmds/pdftexcmds.sty" 1708992705 20089 80423eac55aa175305d35b49e04fe23b ""
"c:/texlive/2023/texmf-dist/tex/generic/uniquecounter/uniquecounter.sty" 1708994449 7008 f92eaa0a3872ed622bbf538217cd2ab7 ""
"c:/texlive/2023/texmf-dist/tex/latex/amscls/amsthm.sty" 1708988587 12594 0d51ac3a545aaaa555021326ff22a6cc ""
"c:/texlive/2023/texmf-dist/tex/latex/amsfonts/amsfonts.sty" 1708988591 5949 3f3fd50a8cc94c3d4cbf4fc66cd3df1c ""
"c:/texlive/2023/texmf-dist/tex/latex/amsfonts/amssymb.sty" 1708988591 13829 94730e64147574077f8ecfea9bb69af4 ""
@@ -44,17 +60,28 @@
"c:/texlive/2023/texmf-dist/tex/latex/amsmath/amsmath.sty" 1708988596 88371 d84032c0f422c3d1e282266c01bef237 ""
"c:/texlive/2023/texmf-dist/tex/latex/amsmath/amsopn.sty" 1708988596 4474 b811654f4bf125f11506d13d13647efb ""
"c:/texlive/2023/texmf-dist/tex/latex/amsmath/amstext.sty" 1708988596 2444 0d0c1ee65478277e8015d65b86983da2 ""
"c:/texlive/2023/texmf-dist/tex/latex/atveryend/atveryend.sty" 1708988739 19336 ce7ae9438967282886b3b036cfad1e4d ""
"c:/texlive/2023/texmf-dist/tex/latex/auxhook/auxhook.sty" 1708988768 3935 57aa3c3e203a5c2effb4d2bd2efbc323 ""
"c:/texlive/2023/texmf-dist/tex/latex/base/article.cls" 1708991500 20144 147463a6a579f4597269ef9565205cfe ""
"c:/texlive/2023/texmf-dist/tex/latex/base/atbegshi-ltx.sty" 1708991500 3045 273c666a54e60b9f730964f431a56c1b ""
"c:/texlive/2023/texmf-dist/tex/latex/base/atveryend-ltx.sty" 1708991500 2462 6bc53756156dbd71c1ad550d30a3b93f ""
"c:/texlive/2023/texmf-dist/tex/latex/base/size11.clo" 1708991500 8464 59874a3b0776c73e2a138b025d8473dd ""
"c:/texlive/2023/texmf-dist/tex/latex/epstopdf-pkg/epstopdf-base.sty" 1708990302 13886 d1306dcf79a944f6988e688c1785f9ce ""
"c:/texlive/2023/texmf-dist/tex/latex/etoolbox/etoolbox.sty" 1708990361 46845 3b58f70c6e861a13d927bff09d35ecbc ""
"c:/texlive/2023/texmf-dist/tex/latex/fancyhdr/fancyhdr.sty" 1708990446 18450 88279bf67c81e69f8e3f1c1bad1a26c5 ""
"c:/texlive/2023/texmf-dist/tex/latex/float/float.sty" 1708990580 6749 16d2656a1984957e674b149555f1ea1d ""
"c:/texlive/2023/texmf-dist/tex/latex/graphics-cfg/graphics.cfg" 1708990878 1224 978390e9c2234eab29404bc21b268d1e ""
"c:/texlive/2023/texmf-dist/tex/latex/graphics-def/pdftex.def" 1708990879 19448 1e988b341dda20961a6b931bcde55519 ""
"c:/texlive/2023/texmf-dist/tex/latex/graphics/graphics.sty" 1708990876 18387 8f900a490197ebaf93c02ae9476d4b09 ""
"c:/texlive/2023/texmf-dist/tex/latex/graphics/graphicx.sty" 1708990876 8010 a8d949cbdbc5c983593827c9eec252e1 ""
"c:/texlive/2023/texmf-dist/tex/latex/graphics/keyval.sty" 1708990876 2671 7e67d78d9b88c845599a85b2d41f2e39 ""
"c:/texlive/2023/texmf-dist/tex/latex/graphics/trig.sty" 1708990876 4023 293ea1c16429fc0c4cf605f4da1791a9 ""
"c:/texlive/2023/texmf-dist/tex/latex/hycolor/hycolor.sty" 1708991092 17914 4c28a13fc3d975e6e81c9bea1d697276 ""
"c:/texlive/2023/texmf-dist/tex/latex/hyperref/hpdftex.def" 1708991102 48154 e46bf8adeb936500541441171d61726d ""
"c:/texlive/2023/texmf-dist/tex/latex/hyperref/hyperref.sty" 1708991102 220920 fd3cbb5f1a2bc9b8f451b8b7d8171264 ""
"c:/texlive/2023/texmf-dist/tex/latex/hyperref/nameref.sty" 1708991102 11026 182c63f139a71afd30a28e5f1ed2cd1c ""
"c:/texlive/2023/texmf-dist/tex/latex/hyperref/pd1enc.def" 1708991102 14249 e67cb186717b7ab18d14a4875e7e98b5 ""
"c:/texlive/2023/texmf-dist/tex/latex/hyperref/puenc.def" 1708991102 117112 05831178ece2cad4d9629dcf65099b11 ""
"c:/texlive/2023/texmf-dist/tex/latex/jknapltx/mathrsfs.sty" 1708991316 300 12fa6f636b617656f2810ee82cb05015 ""
"c:/texlive/2023/texmf-dist/tex/latex/jknapltx/ursfs.fd" 1708991316 548 cc4e3557704bfed27c7002773fad6c90 ""
"c:/texlive/2023/texmf-dist/tex/latex/kvoptions/kvoptions.sty" 1708991458 22555 6d8e155cfef6d82c3d5c742fea7c992e ""
@@ -65,16 +92,28 @@
"c:/texlive/2023/texmf-dist/tex/latex/mathtools/mhsetup.sty" 1708991994 5582 a43dedf8e5ec418356f1e9dfe5d29fc3 ""
"c:/texlive/2023/texmf-dist/tex/latex/parskip/parskip.sty" 1708992628 4288 94714aa7f535440f33181fec52a31963 ""
"c:/texlive/2023/texmf-dist/tex/latex/preprint/fullpage.sty" 1708992914 2789 05b418f78b224ec872f5b11081138605 ""
"c:/texlive/2023/texmf-dist/tex/latex/refcount/refcount.sty" 1708993268 9878 9e94e8fa600d95f9c7731bb21dfb67a4 ""
"c:/texlive/2023/texmf-dist/tex/latex/rerunfilecheck/rerunfilecheck.sty" 1708993294 9714 ba3194bd52c8499b3f1e3eb91d409670 ""
"c:/texlive/2023/texmf-dist/tex/latex/tools/calc.sty" 1708994243 10214 547fd4d29642cb7c80bf54b49d447f01 ""
"c:/texlive/2023/texmf-dist/tex/latex/url/url.sty" 1708994494 12796 8edb7d69a20b857904dd0ea757c14ec9 ""
"c:/texlive/2023/texmf-dist/web2c/texmf.cnf" 1708988443 41009 84b61f42d16d06bedb915f57aa2374cf ""
"c:/texlive/2023/texmf-var/fonts/map/pdftex/updmap/pdftex.map" 1708994999 5518052 de2a91c664d75f3971de4662dc6b5a65 ""
"c:/texlive/2023/texmf-var/web2c/pdftex/pdflatex.fmt" 1708995327 8220658 fb4d14532342a0ef5245dd396c4a1bd1 ""
"c:/texlive/2023/texmf.cnf" 1708994944 713 e69b156964470283e0530f5060668171 ""
"d:/Documents/Nextcloud/Documents/Project WUSTL/Academic/2025_Fall/CSE5100/Homeworks/hw2/result.tex" 1760230163 6257 29867be9781c52dc4faad49bb0cba6fa ""
"result.aux" 1760230165 119 495be67432001ea8f4f9fa642ad39ad3 "pdflatex"
"result.tex" 1760230163 6257 29867be9781c52dc4faad49bb0cba6fa ""
"d:/Documents/Nextcloud/Documents/Project WUSTL/Academic/2025_Fall/CSE5100/Homeworks/hw2/result.tex" 1760491969 10252 4b623ef8b8c8e01ee79b1be0d47dadd2 ""
"result.aux" 1760491972 1925 503142906fd718693c8f6c9386c260f9 "pdflatex"
"result.out" 1760491971 0 d41d8cd98f00b204e9800998ecf8427e "pdflatex"
"result.tex" 1760491969 10252 4b623ef8b8c8e01ee79b1be0d47dadd2 ""
"runs/DQN/results.png" 1760413579 78748 522d4ff2da0cc1db482579c8f93c9fb2 ""
"runs/Double DQN/results.png" 1760421509 83905 dc93c0d523de8b27e35ade0e21f0972c ""
"runs/Dueling DQN/results.png" 1760465274 81080 e588bde8f39161b815b9d842a32edffe ""
"runs/NStep + PER/results.png" 1760488208 72434 d0141a2aad159aacd350d13606ca6e5b ""
"runs/NStep/results.png" 1760487001 68663 604fd7f3421f7652e124f62849e4916b ""
"runs/Noisy DQN/results.png" 1760491128 85111 706ee9849d7367ed9f2d467ac6284665 ""
"runs/PER/results.png" 1760466054 86443 5c386bc256bcc4c871bbf8d8ae48b943 ""
(generated)
"result.aux"
"result.log"
"result.out"
"result.pdf"
(rewritten before read)

View File

@@ -51,6 +51,57 @@ INPUT c:/texlive/2023/texmf-dist/tex/latex/tools/calc.sty
INPUT c:/texlive/2023/texmf-dist/tex/latex/tools/calc.sty
INPUT c:/texlive/2023/texmf-dist/tex/latex/mathtools/mhsetup.sty
INPUT c:/texlive/2023/texmf-dist/tex/latex/mathtools/mhsetup.sty
INPUT c:/texlive/2023/texmf-dist/tex/latex/float/float.sty
INPUT c:/texlive/2023/texmf-dist/tex/latex/float/float.sty
INPUT c:/texlive/2023/texmf-dist/tex/latex/hyperref/hyperref.sty
INPUT c:/texlive/2023/texmf-dist/tex/latex/hyperref/hyperref.sty
INPUT c:/texlive/2023/texmf-dist/tex/generic/iftex/iftex.sty
INPUT c:/texlive/2023/texmf-dist/tex/generic/iftex/iftex.sty
INPUT c:/texlive/2023/texmf-dist/tex/generic/kvdefinekeys/kvdefinekeys.sty
INPUT c:/texlive/2023/texmf-dist/tex/generic/kvdefinekeys/kvdefinekeys.sty
INPUT c:/texlive/2023/texmf-dist/tex/generic/pdfescape/pdfescape.sty
INPUT c:/texlive/2023/texmf-dist/tex/generic/pdfescape/pdfescape.sty
INPUT c:/texlive/2023/texmf-dist/tex/generic/pdftexcmds/pdftexcmds.sty
INPUT c:/texlive/2023/texmf-dist/tex/generic/pdftexcmds/pdftexcmds.sty
INPUT c:/texlive/2023/texmf-dist/tex/generic/infwarerr/infwarerr.sty
INPUT c:/texlive/2023/texmf-dist/tex/generic/infwarerr/infwarerr.sty
INPUT c:/texlive/2023/texmf-dist/tex/latex/hycolor/hycolor.sty
INPUT c:/texlive/2023/texmf-dist/tex/latex/hycolor/hycolor.sty
INPUT c:/texlive/2023/texmf-dist/tex/latex/auxhook/auxhook.sty
INPUT c:/texlive/2023/texmf-dist/tex/latex/auxhook/auxhook.sty
INPUT c:/texlive/2023/texmf-dist/tex/latex/hyperref/nameref.sty
INPUT c:/texlive/2023/texmf-dist/tex/latex/hyperref/nameref.sty
INPUT c:/texlive/2023/texmf-dist/tex/latex/refcount/refcount.sty
INPUT c:/texlive/2023/texmf-dist/tex/latex/refcount/refcount.sty
INPUT c:/texlive/2023/texmf-dist/tex/generic/gettitlestring/gettitlestring.sty
INPUT c:/texlive/2023/texmf-dist/tex/generic/gettitlestring/gettitlestring.sty
INPUT c:/texlive/2023/texmf-dist/tex/latex/hyperref/pd1enc.def
INPUT c:/texlive/2023/texmf-dist/tex/latex/hyperref/pd1enc.def
INPUT c:/texlive/2023/texmf-dist/tex/latex/hyperref/pd1enc.def
INPUT c:/texlive/2023/texmf-dist/tex/generic/intcalc/intcalc.sty
INPUT c:/texlive/2023/texmf-dist/tex/generic/intcalc/intcalc.sty
INPUT c:/texlive/2023/texmf-dist/tex/latex/hyperref/puenc.def
INPUT c:/texlive/2023/texmf-dist/tex/latex/hyperref/puenc.def
INPUT c:/texlive/2023/texmf-dist/tex/latex/hyperref/puenc.def
INPUT c:/texlive/2023/texmf-dist/tex/latex/url/url.sty
INPUT c:/texlive/2023/texmf-dist/tex/latex/url/url.sty
INPUT c:/texlive/2023/texmf-dist/tex/generic/bitset/bitset.sty
INPUT c:/texlive/2023/texmf-dist/tex/generic/bitset/bitset.sty
INPUT c:/texlive/2023/texmf-dist/tex/generic/bigintcalc/bigintcalc.sty
INPUT c:/texlive/2023/texmf-dist/tex/generic/bigintcalc/bigintcalc.sty
INPUT c:/texlive/2023/texmf-dist/tex/generic/atbegshi/atbegshi.sty
INPUT c:/texlive/2023/texmf-dist/tex/latex/base/atbegshi-ltx.sty
INPUT c:/texlive/2023/texmf-dist/tex/latex/base/atbegshi-ltx.sty
INPUT c:/texlive/2023/texmf-dist/tex/latex/hyperref/hpdftex.def
INPUT c:/texlive/2023/texmf-dist/tex/latex/hyperref/hpdftex.def
INPUT c:/texlive/2023/texmf-dist/tex/latex/hyperref/hpdftex.def
INPUT c:/texlive/2023/texmf-dist/tex/latex/atveryend/atveryend.sty
INPUT c:/texlive/2023/texmf-dist/tex/latex/base/atveryend-ltx.sty
INPUT c:/texlive/2023/texmf-dist/tex/latex/base/atveryend-ltx.sty
INPUT c:/texlive/2023/texmf-dist/tex/latex/rerunfilecheck/rerunfilecheck.sty
INPUT c:/texlive/2023/texmf-dist/tex/latex/rerunfilecheck/rerunfilecheck.sty
INPUT c:/texlive/2023/texmf-dist/tex/generic/uniquecounter/uniquecounter.sty
INPUT c:/texlive/2023/texmf-dist/tex/generic/uniquecounter/uniquecounter.sty
INPUT c:/texlive/2023/texmf-dist/tex/latex/l3backend/l3backend-pdftex.def
INPUT c:/texlive/2023/texmf-dist/tex/latex/l3backend/l3backend-pdftex.def
INPUT ./result.aux
@@ -77,7 +128,16 @@ INPUT c:/texlive/2023/texmf-dist/tex/latex/epstopdf-pkg/epstopdf-base.sty
INPUT c:/texlive/2023/texmf-dist/tex/latex/latexconfig/epstopdf-sys.cfg
INPUT c:/texlive/2023/texmf-dist/tex/latex/latexconfig/epstopdf-sys.cfg
INPUT c:/texlive/2023/texmf-dist/tex/latex/latexconfig/epstopdf-sys.cfg
INPUT ./result.out
INPUT ./result.out
INPUT result.out
INPUT result.out
INPUT ./result.out
INPUT ./result.out
OUTPUT result.out
OUTPUT result.pdf
INPUT c:/texlive/2023/texmf-dist/fonts/tfm/public/cm/cmbx10.tfm
INPUT c:/texlive/2023/texmf-dist/fonts/tfm/public/cm/cmtt10.tfm
INPUT c:/texlive/2023/texmf-dist/fonts/tfm/public/cm/cmr8.tfm
INPUT c:/texlive/2023/texmf-dist/fonts/tfm/public/cm/cmr6.tfm
INPUT c:/texlive/2023/texmf-dist/fonts/tfm/public/cm/cmmi10.tfm
@@ -107,10 +167,48 @@ INPUT c:/texlive/2023/texmf-dist/tex/latex/jknapltx/ursfs.fd
INPUT c:/texlive/2023/texmf-dist/fonts/tfm/public/rsfs/rsfs10.tfm
INPUT c:/texlive/2023/texmf-dist/fonts/tfm/public/rsfs/rsfs10.tfm
INPUT c:/texlive/2023/texmf-dist/fonts/tfm/public/rsfs/rsfs5.tfm
INPUT c:/texlive/2023/texmf-dist/fonts/tfm/public/cm/cmti10.tfm
OUTPUT result.pdf
INPUT c:/texlive/2023/texmf-dist/fonts/tfm/jknappen/ec/tcrm1095.tfm
INPUT c:/texlive/2023/texmf-var/fonts/map/pdftex/updmap/pdftex.map
INPUT c:/texlive/2023/texmf-dist/fonts/enc/dvips/cm-super/cm-super-ts1.enc
INPUT c:/texlive/2023/texmf-dist/fonts/tfm/public/cm/cmti10.tfm
INPUT ./runs/DQN/results.png
INPUT ./runs/DQN/results.png
INPUT ./runs/DQN/results.png
INPUT ./runs/DQN/results.png
INPUT ./runs/DQN/results.png
INPUT ./runs/Double DQN/results.png
INPUT ./runs/Double DQN/results.png
INPUT ./runs/Double DQN/results.png
INPUT ./runs/Double DQN/results.png
INPUT ./runs/Double DQN/results.png
INPUT ./runs/Dueling DQN/results.png
INPUT ./runs/Dueling DQN/results.png
INPUT ./runs/Dueling DQN/results.png
INPUT ./runs/Dueling DQN/results.png
INPUT ./runs/Dueling DQN/results.png
INPUT ./runs/PER/results.png
INPUT ./runs/PER/results.png
INPUT ./runs/PER/results.png
INPUT ./runs/PER/results.png
INPUT ./runs/PER/results.png
INPUT ./runs/NStep/results.png
INPUT ./runs/NStep/results.png
INPUT ./runs/NStep/results.png
INPUT ./runs/NStep/results.png
INPUT ./runs/NStep/results.png
INPUT ./runs/NStep + PER/results.png
INPUT ./runs/NStep + PER/results.png
INPUT ./runs/NStep + PER/results.png
INPUT ./runs/NStep + PER/results.png
INPUT ./runs/NStep + PER/results.png
INPUT ./runs/Noisy DQN/results.png
INPUT ./runs/Noisy DQN/results.png
INPUT ./runs/Noisy DQN/results.png
INPUT ./runs/Noisy DQN/results.png
INPUT ./runs/Noisy DQN/results.png
INPUT result.aux
INPUT ./result.out
INPUT ./result.out
INPUT c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmbx10.pfb
INPUT c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmex10.pfb
INPUT c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmmi10.pfb
@@ -121,4 +219,6 @@ INPUT c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmsy10.pfb
INPUT c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmsy6.pfb
INPUT c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmsy8.pfb
INPUT c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmti10.pfb
INPUT c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmtt10.pfb
INPUT c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/symbols/msbm10.pfb
INPUT c:/texlive/2023/texmf-dist/fonts/type1/public/cm-super/sfrm1095.pfb

View File

@@ -1,4 +1,4 @@
This is pdfTeX, Version 3.141592653-2.6-1.40.25 (TeX Live 2023) (preloaded format=pdflatex 2024.2.26) 11 OCT 2025 19:49
This is pdfTeX, Version 3.141592653-2.6-1.40.25 (TeX Live 2023) (preloaded format=pdflatex 2024.2.26) 14 OCT 2025 20:32
entering extended mode
restricted \write18 enabled.
file:line:error style messages enabled.
@@ -166,29 +166,139 @@ Package: mhsetup 2021/03/18 v1.4 programming setup (MH)
\l_MT_below_shortintertext_sep=\dimen158
\xmathstrut@box=\box53
\xmathstrut@dim=\dimen159
) (c:/texlive/2023/texmf-dist/tex/latex/float/float.sty
Package: float 2001/11/08 v1.3d Float enhancements (AL)
\c@float@type=\count279
\float@exts=\toks29
\float@box=\box54
\@float@everytoks=\toks30
\@floatcapt=\box55
) (c:/texlive/2023/texmf-dist/tex/latex/hyperref/hyperref.sty
Package: hyperref 2024-01-20 v7.01h Hypertext links for LaTeX
(c:/texlive/2023/texmf-dist/tex/generic/iftex/iftex.sty
Package: iftex 2022/02/03 v1.0f TeX engine tests
) (c:/texlive/2023/texmf-dist/tex/generic/kvdefinekeys/kvdefinekeys.sty
Package: kvdefinekeys 2019-12-19 v1.6 Define keys (HO)
) (c:/texlive/2023/texmf-dist/tex/generic/pdfescape/pdfescape.sty
Package: pdfescape 2019/12/09 v1.15 Implements pdfTeX's escape features (HO)
(c:/texlive/2023/texmf-dist/tex/generic/pdftexcmds/pdftexcmds.sty
Package: pdftexcmds 2020-06-27 v0.33 Utility functions of pdfTeX for LuaTeX (HO)
(c:/texlive/2023/texmf-dist/tex/generic/infwarerr/infwarerr.sty
Package: infwarerr 2019/12/03 v1.5 Providing info/warning/error messages (HO)
)
\c@theorem=\count279
Package pdftexcmds Info: \pdf@primitive is available.
Package pdftexcmds Info: \pdf@ifprimitive is available.
Package pdftexcmds Info: \pdfdraftmode found.
)) (c:/texlive/2023/texmf-dist/tex/latex/hycolor/hycolor.sty
Package: hycolor 2020-01-27 v1.10 Color options for hyperref/bookmark (HO)
) (c:/texlive/2023/texmf-dist/tex/latex/auxhook/auxhook.sty
Package: auxhook 2019-12-17 v1.6 Hooks for auxiliary files (HO)
) (c:/texlive/2023/texmf-dist/tex/latex/hyperref/nameref.sty
Package: nameref 2023-11-26 v2.56 Cross-referencing by name of section
(c:/texlive/2023/texmf-dist/tex/latex/refcount/refcount.sty
Package: refcount 2019/12/15 v3.6 Data extraction from label references (HO)
) (c:/texlive/2023/texmf-dist/tex/generic/gettitlestring/gettitlestring.sty
Package: gettitlestring 2019/12/15 v1.6 Cleanup title references (HO)
)
\c@section@level=\count280
)
\@linkdim=\dimen160
\Hy@linkcounter=\count281
\Hy@pagecounter=\count282
(c:/texlive/2023/texmf-dist/tex/latex/hyperref/pd1enc.def
File: pd1enc.def 2024-01-20 v7.01h Hyperref: PDFDocEncoding definition (HO)
Now handling font encoding PD1 ...
... no UTF-8 mapping file for font encoding PD1
) (c:/texlive/2023/texmf-dist/tex/generic/intcalc/intcalc.sty
Package: intcalc 2019/12/15 v1.3 Expandable calculations with integers (HO)
)
\Hy@SavedSpaceFactor=\count283
(c:/texlive/2023/texmf-dist/tex/latex/hyperref/puenc.def
File: puenc.def 2024-01-20 v7.01h Hyperref: PDF Unicode definition (HO)
Now handling font encoding PU ...
... no UTF-8 mapping file for font encoding PU
)
Package hyperref Info: Hyper figures OFF on input line 4179.
Package hyperref Info: Link nesting OFF on input line 4184.
Package hyperref Info: Hyper index ON on input line 4187.
Package hyperref Info: Plain pages OFF on input line 4194.
Package hyperref Info: Backreferencing OFF on input line 4199.
Package hyperref Info: Implicit mode ON; LaTeX internals redefined.
Package hyperref Info: Bookmarks ON on input line 4446.
\c@Hy@tempcnt=\count284
(c:/texlive/2023/texmf-dist/tex/latex/url/url.sty
\Urlmuskip=\muskip17
Package: url 2013/09/16 ver 3.4 Verb mode for urls, etc.
)
LaTeX Info: Redefining \url on input line 4784.
\XeTeXLinkMargin=\dimen161
(c:/texlive/2023/texmf-dist/tex/generic/bitset/bitset.sty
Package: bitset 2019/12/09 v1.3 Handle bit-vector datatype (HO)
(c:/texlive/2023/texmf-dist/tex/generic/bigintcalc/bigintcalc.sty
Package: bigintcalc 2019/12/15 v1.5 Expandable calculations on big integers (HO)
))
\Fld@menulength=\count285
\Field@Width=\dimen162
\Fld@charsize=\dimen163
Package hyperref Info: Hyper figures OFF on input line 6063.
Package hyperref Info: Link nesting OFF on input line 6068.
Package hyperref Info: Hyper index ON on input line 6071.
Package hyperref Info: backreferencing OFF on input line 6078.
Package hyperref Info: Link coloring OFF on input line 6083.
Package hyperref Info: Link coloring with OCG OFF on input line 6088.
Package hyperref Info: PDF/A mode OFF on input line 6093.
(c:/texlive/2023/texmf-dist/tex/latex/base/atbegshi-ltx.sty
Package: atbegshi-ltx 2021/01/10 v1.0c Emulation of the original atbegshi
package with kernel methods
)
\Hy@abspage=\count286
\c@Item=\count287
\c@Hfootnote=\count288
)
Package hyperref Info: Driver (autodetected): hpdftex.
(c:/texlive/2023/texmf-dist/tex/latex/hyperref/hpdftex.def
File: hpdftex.def 2024-01-20 v7.01h Hyperref driver for pdfTeX
(c:/texlive/2023/texmf-dist/tex/latex/base/atveryend-ltx.sty
Package: atveryend-ltx 2020/08/19 v1.0a Emulation of the original atveryend package
with kernel methods
)
\Fld@listcount=\count289
\c@bookmark@seq@number=\count290
(c:/texlive/2023/texmf-dist/tex/latex/rerunfilecheck/rerunfilecheck.sty
Package: rerunfilecheck 2022-07-10 v1.10 Rerun checks for auxiliary files (HO)
(c:/texlive/2023/texmf-dist/tex/generic/uniquecounter/uniquecounter.sty
Package: uniquecounter 2019/12/15 v1.4 Provide unlimited unique counter (HO)
)
Package uniquecounter Info: New unique counter `rerunfilecheck' on input line 285.
)
\Hy@SectionHShift=\skip70
)
\c@theorem=\count291
(c:/texlive/2023/texmf-dist/tex/latex/l3backend/l3backend-pdftex.def
File: l3backend-pdftex.def 2024-02-20 L3 backend support: PDF output (pdfTeX)
\l__color_backend_stack_int=\count280
\l__pdf_internal_box=\box54
\l__color_backend_stack_int=\count292
\l__pdf_internal_box=\box56
) (./result.aux)
\openout1 = `result.aux'.
LaTeX Font Info: Checking defaults for OML/cmm/m/it on input line 77.
LaTeX Font Info: ... okay on input line 77.
LaTeX Font Info: Checking defaults for OMS/cmsy/m/n on input line 77.
LaTeX Font Info: ... okay on input line 77.
LaTeX Font Info: Checking defaults for OT1/cmr/m/n on input line 77.
LaTeX Font Info: ... okay on input line 77.
LaTeX Font Info: Checking defaults for T1/cmr/m/n on input line 77.
LaTeX Font Info: ... okay on input line 77.
LaTeX Font Info: Checking defaults for TS1/cmr/m/n on input line 77.
LaTeX Font Info: ... okay on input line 77.
LaTeX Font Info: Checking defaults for OMX/cmex/m/n on input line 77.
LaTeX Font Info: ... okay on input line 77.
LaTeX Font Info: Checking defaults for U/cmr/m/n on input line 77.
LaTeX Font Info: ... okay on input line 77.
LaTeX Font Info: Checking defaults for OML/cmm/m/it on input line 79.
LaTeX Font Info: ... okay on input line 79.
LaTeX Font Info: Checking defaults for OMS/cmsy/m/n on input line 79.
LaTeX Font Info: ... okay on input line 79.
LaTeX Font Info: Checking defaults for OT1/cmr/m/n on input line 79.
LaTeX Font Info: ... okay on input line 79.
LaTeX Font Info: Checking defaults for T1/cmr/m/n on input line 79.
LaTeX Font Info: ... okay on input line 79.
LaTeX Font Info: Checking defaults for TS1/cmr/m/n on input line 79.
LaTeX Font Info: ... okay on input line 79.
LaTeX Font Info: Checking defaults for OMX/cmex/m/n on input line 79.
LaTeX Font Info: ... okay on input line 79.
LaTeX Font Info: Checking defaults for U/cmr/m/n on input line 79.
LaTeX Font Info: ... okay on input line 79.
LaTeX Font Info: Checking defaults for PD1/pdf/m/n on input line 79.
LaTeX Font Info: ... okay on input line 79.
LaTeX Font Info: Checking defaults for PU/pdf/m/n on input line 79.
LaTeX Font Info: ... okay on input line 79.
(c:/texlive/2023/texmf-dist/tex/latex/graphics/graphicx.sty
Package: graphicx 2021/09/16 v1.2d Enhanced LaTeX Graphics (DPC,SPQR)
(c:/texlive/2023/texmf-dist/tex/latex/graphics/graphics.sty
@@ -203,58 +313,108 @@ Package graphics Info: Driver file: pdftex.def on input line 107.
File: pdftex.def 2022/09/22 v1.2b Graphics/color driver for pdftex
(c:/texlive/2023/texmf-dist/tex/context/base/mkii/supp-pdf.mkii
[Loading MPS to PDF converter (version 2006.09.02).]
\scratchcounter=\count281
\scratchdimen=\dimen160
\scratchbox=\box55
\nofMPsegments=\count282
\nofMParguments=\count283
\everyMPshowfont=\toks29
\MPscratchCnt=\count284
\MPscratchDim=\dimen161
\MPnumerator=\count285
\makeMPintoPDFobject=\count286
\everyMPtoPDFconversion=\toks30
\scratchcounter=\count293
\scratchdimen=\dimen164
\scratchbox=\box57
\nofMPsegments=\count294
\nofMParguments=\count295
\everyMPshowfont=\toks31
\MPscratchCnt=\count296
\MPscratchDim=\dimen165
\MPnumerator=\count297
\makeMPintoPDFobject=\count298
\everyMPtoPDFconversion=\toks32
))) (c:/texlive/2023/texmf-dist/tex/latex/epstopdf-pkg/epstopdf-base.sty
Package: epstopdf-base 2020-01-24 v2.11 Base part for package epstopdf
Package epstopdf-base Info: Redefining graphics rule for `.eps' on input line 485.
(c:/texlive/2023/texmf-dist/tex/latex/latexconfig/epstopdf-sys.cfg
File: epstopdf-sys.cfg 2010/07/13 v1.3 Configuration of (r)epstopdf for TeX Live
))
\Gin@req@height=\dimen162
\Gin@req@width=\dimen163
\Gin@req@height=\dimen166
\Gin@req@width=\dimen167
)
LaTeX Font Info: Trying to load font information for U+msa on input line 88.
Package hyperref Info: Link coloring OFF on input line 79.
(./result.out) (./result.out)
\@outlinefile=\write3
\openout3 = `result.out'.
LaTeX Font Info: Trying to load font information for U+msa on input line 92.
(c:/texlive/2023/texmf-dist/tex/latex/amsfonts/umsa.fd
File: umsa.fd 2013/01/14 v3.01 AMS symbols A
)
LaTeX Font Info: Trying to load font information for U+msb on input line 88.
LaTeX Font Info: Trying to load font information for U+msb on input line 92.
(c:/texlive/2023/texmf-dist/tex/latex/amsfonts/umsb.fd
File: umsb.fd 2013/01/14 v3.01 AMS symbols B
)
LaTeX Font Info: Trying to load font information for U+rsfs on input line 88.
LaTeX Font Info: Trying to load font information for U+rsfs on input line 92.
(c:/texlive/2023/texmf-dist/tex/latex/jknapltx/ursfs.fd
File: ursfs.fd 1998/03/24 rsfs font definition file (jk)
) [1{c:/texlive/2023/texmf-var/fonts/map/pdftex/updmap/pdftex.map}
)
Overfull \hbox (1.79282pt too wide) in paragraph at lines 92--93
[]\OT1/cmr/m/n/10.95 This home-work is com-pleted with the help of Wind-surf VS code ex-ten-sion.[]$\OT1/cmtt/m/n/10.95 https : / / windsurf .
[]
] [2] [3] [4] (./result.aux)
[1{c:/texlive/2023/texmf-var/fonts/map/pdftex/updmap/pdftex.map}
{c:/texlive/2023/texmf-dist/fonts/enc/dvips/cm-super/cm-super-ts1.enc}] [2] [3]
<./runs/DQN/results.png, id=40, 1445.4pt x 433.62pt>
File: ./runs/DQN/results.png Graphic file (type png)
<use ./runs/DQN/results.png>
Package pdftex.def Info: ./runs/DQN/results.png used on input line 178.
(pdftex.def) Requested size: 422.77664pt x 126.83168pt.
<./runs/Double DQN/results.png, id=42, 1445.4pt x 433.62pt>
File: ./runs/Double DQN/results.png Graphic file (type png)
<use ./runs/Double DQN/results.png>
Package pdftex.def Info: ./runs/Double DQN/results.png used on input line 186.
(pdftex.def) Requested size: 422.77664pt x 126.83168pt.
<./runs/Dueling DQN/results.png, id=43, 1445.4pt x 433.62pt>
File: ./runs/Dueling DQN/results.png Graphic file (type png)
<use ./runs/Dueling DQN/results.png>
Package pdftex.def Info: ./runs/Dueling DQN/results.png used on input line 194.
(pdftex.def) Requested size: 422.77664pt x 126.83168pt.
<./runs/PER/results.png, id=44, 1445.4pt x 433.62pt>
File: ./runs/PER/results.png Graphic file (type png)
<use ./runs/PER/results.png>
Package pdftex.def Info: ./runs/PER/results.png used on input line 202.
(pdftex.def) Requested size: 469.75502pt x 140.92482pt.
[4 <./runs/DQN/results.png> <./runs/Double DQN/results.png> <./runs/Dueling DQN/results.png>]
<./runs/NStep/results.png, id=55, 1445.4pt x 433.62pt>
File: ./runs/NStep/results.png Graphic file (type png)
<use ./runs/NStep/results.png>
Package pdftex.def Info: ./runs/NStep/results.png used on input line 210.
(pdftex.def) Requested size: 469.75502pt x 140.92482pt.
<./runs/NStep + PER/results.png, id=56, 1445.4pt x 433.62pt>
File: ./runs/NStep + PER/results.png Graphic file (type png)
<use ./runs/NStep + PER/results.png>
Package pdftex.def Info: ./runs/NStep + PER/results.png used on input line 218.
(pdftex.def) Requested size: 469.75502pt x 140.92482pt.
[5 <./runs/PER/results.png> <./runs/NStep/results.png>]
<./runs/Noisy DQN/results.png, id=65, 1445.4pt x 433.62pt>
File: ./runs/Noisy DQN/results.png Graphic file (type png)
<use ./runs/Noisy DQN/results.png>
Package pdftex.def Info: ./runs/Noisy DQN/results.png used on input line 226.
(pdftex.def) Requested size: 469.75502pt x 140.92482pt.
[6 <./runs/NStep + PER/results.png> <./runs/Noisy DQN/results.png>] [7] (./result.aux)
***********
LaTeX2e <2023-11-01> patch level 1
L3 programming layer <2024-02-20>
***********
Package rerunfilecheck Info: File `result.out' has not changed.
(rerunfilecheck) Checksum: D41D8CD98F00B204E9800998ECF8427E;0.
)
Here is how much of TeX's memory you used:
4245 strings out of 474137
64387 string characters out of 5748517
1938190 words of memory out of 5000000
26537 multiletter control sequences out of 15000+600000
563865 words of font info for 59 fonts, out of 8000000 for 9000
10704 strings out of 474137
164550 string characters out of 5748517
1940190 words of memory out of 5000000
32881 multiletter control sequences out of 15000+600000
564374 words of font info for 61 fonts, out of 8000000 for 9000
1141 hyphenation exceptions out of 8191
65i,11n,72p,713b,463s stack positions out of 10000i,1000n,20000p,200000b,200000s
<c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmbx10.pfb><c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmex10.pfb><c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmmi10.pfb><c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmmi8.pfb><c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmr10.pfb><c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmr8.pfb><c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmsy10.pfb><c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmsy6.pfb><c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmsy8.pfb><c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmti10.pfb><c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/symbols/msbm10.pfb>
Output written on result.pdf (4 pages, 122012 bytes).
69i,11n,79p,713b,615s stack positions out of 10000i,1000n,20000p,200000b,200000s
<c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmbx10.pfb><c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmex10.pfb><c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmmi10.pfb><c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmmi8.pfb><c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmr10.pfb><c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmr8.pfb><c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmsy10.pfb><c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmsy6.pfb><c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmsy8.pfb><c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmti10.pfb><c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/cm/cmtt10.pfb><c:/texlive/2023/texmf-dist/fonts/type1/public/amsfonts/symbols/msbm10.pfb><c:/texlive/2023/texmf-dist/fonts/type1/public/cm-super/sfrm1095.pfb>
Output written on result.pdf (7 pages, 678475 bytes).
PDF statistics:
72 PDF objects out of 1000 (max. 8388607)
43 compressed objects within 1 object stream
0 named destinations out of 1000 (max. 500000)
1 words of extra memory for PDF output out of 10000 (max. 10000000)
141 PDF objects out of 1000 (max. 8388607)
91 compressed objects within 1 object stream
21 named destinations out of 1000 (max. 500000)
36 words of extra memory for PDF output out of 10000 (max. 10000000)

Binary file not shown.

Binary file not shown.

View File

@@ -5,6 +5,8 @@
\usepackage{fullpage}
\usepackage{mathrsfs}
\usepackage{mathtools}
\usepackage{float}
\usepackage{hyperref}
%%
%% Stuff above here is packages that will be used to compile your document.
@@ -85,6 +87,29 @@
\begin{enumerate}
\item[] \textbf{Use Of GenAI}
This homework is completed with the help of Windsurf VS code extension.\url{https://windsurf.com/}
What is used:
\begin{itemize}
\item Autofill feature to generate syntactically correct latex code (each tab key pressed filled no more than 100 characters, at most $20\%$ of the predicted text is adapted) for the homework with human supervision.
\item Use AI to debug the latex code and find unclosed parentheses or other syntax errors.
\item Use AI to autofill the parts that follows the same structure as the previous parts (example: case by case proofs).
\item Use AI to auto correct misspelled words or latex commands.
\end{itemize}
What is not used:
\begin{itemize}
\item Directly use AI to generate the solutions in latex document.
\item Use AI to ask for hint or solution for the problems.
\item Select part of the document and ask AI to fill the parts missing.
\end{itemize}
\newpage
\item[1.] \textbf{Answer questions in Section 3} Due to the state space complexity of some visual input environments, we may represent Q-functions using a class of parameterized function approximators $\mathcal{Q}=\{Q_w\mid w\in \R^p\}$, where $p$ is the number of parameters. Remember that in the \textit{tabular setting} given a 4-tuple of sampled experience $(s,a,r,s')$, the vanilla Q-learning update is
\[
@@ -109,12 +134,14 @@ where the dependency of $\max_{a'\in A} Q_w(s',a')$ on $w$ is ignored, i.e., it
\item [1.] [\textbf{10pt}] Show that the update \ref{1} and update \ref{2} are the same when the functions in $\mathcal{Q}$ are of the form $Q_w(s,a)=w^T\phi(s,a)$, with $w\in \R^{|S||A|}$ and $\phi:S\times A\to \R^{|S||A|}$, where the feature function $\phi$ is of the form $\phi(s,a)_{s',a'}=\mathbb{I}[s'=s,a'=a]$, where $\mathbb{I}$ denotes the indicator function which evaluates to $1$ if the condition evaluates to true and vice versa. Note that the coordinates in the vector space $\R^{|S||A|}$ can be seen as being indexed by pairs $(s',a')$, where $s'\in S$, $a'\in A$.
\begin{proof}
When the functions in $\mathcal{Q}$ are of the form $Q_w(s,a)=w^T\phi(s,a)$, with $w\in \R^{|S||A|}$ and $\phi:S\times A\to \R^{|S||A|}$, then it is linear.
When the functions in $\mathcal{Q}$ are of the form $Q_w(s,a)=w^T\phi(s,a)$, with $w\in \R^{|S||A|}$ and $\phi:S\times A\to \R^{|S||A|}$, note that $\sum_{s\in S}\sum_{a\in A} \phi(s,a)^T\phi(s,a)=\sum_{s\in S}\sum_{a\in A} \mathbb{I}[s'=s,a'=a]=1$.
\[
\begin{aligned}
Q(s,a)&= Q(s,a)+\alpha\left(r+\gamma\max_{a'\in A} Q(s',a')-Q(s,a)\right)\\
w^T\phi(s,a)&= w^T\phi(s,a)+\alpha\left(r+\gamma\max_{a'\in A} Q(s',a')-Q(s,a)\right)\\
w^T\phi(s,a)&= w^T\phi(s,a)+\alpha\left(r+\gamma\max_{a'\in A} Q(s',a')-Q(s,a)\right)\phi(s,a)^T\phi(s,a)\\
w^T\phi(s,a)&= w^T\phi(s,a)+\alpha\left(r+\gamma\max_{a'\in A} Q(s',a')-Q(s,a)\right)\nabla_w (w^T\phi(s,a))^T\phi(s,a)\\
w^T\phi(s,a)&=\left(w^T+\alpha\left(r+\gamma\max_{a'\in A} Q(s',a')-Q(s,a)\right)\nabla_w Q_w(s,a)\right)^T\phi(s,a)\\
w&= w+\alpha\left(r+\gamma\max_{a'\in A} Q(s',a')-Q(s,a)\right)\nabla_w Q_w(s,a)
\end{aligned}
\]
@@ -143,10 +170,71 @@ where the dependency of $\max_{a'\in A} Q_w(s',a')$ on $w$ is ignored, i.e., it
\item [2.] \textbf{The auto-generated results figure} along with a brief description about what has the figures shown.
\begin{enumerate}
\item [1.] \textbf{DQN}
\begin{figure}[H]
\centering
\includegraphics[width=0.9\textwidth]{./runs/DQN/results.png}
\caption{DQN. Nothing to say but what expected from training.}
\end{figure}
\item [2.] \textbf{Double DQN}
\begin{figure}[H]
\centering
\includegraphics[width=0.9\textwidth]{./runs/Double DQN/results.png}
\caption{Double DQN. I found there is interesting camel like bump for q-value when training with Double DQN. It is less stable than the vanilla DQN.}
\end{figure}
\item [3.] \textbf{Dueling DQN}
\begin{figure}[H]
\centering
\includegraphics[width=0.9\textwidth]{./runs/Dueling DQN/results.png}
\caption{Dueling DQN. Using Advantage network creates comparable results as the DQN.}
\end{figure}
\item [4.] \textbf{Prioritized Experience Replay}
\begin{figure}[H]
\centering
\includegraphics[width=1.0\textwidth]{./runs/PER/results.png}
\caption{Prioritized Experience Replay. Using this alone makes the training process less stable and loss is significantly higher than the previous methods.}
\end{figure}
\item [5.] \textbf{N-Step Experience Replay}
\begin{figure}[H]
\centering
\includegraphics[width=1.0\textwidth]{./runs/NStep/results.png}
\caption{N-Step Experience Replay. So far the most stable method of training, especially when the replay buffer size is large. However, when the replay buffer size is too small, typically $\le 70$, the training process may not converge to optimal performance.}
\end{figure}
\item [6.] \textbf{N-Step + PER}
\begin{figure}[H]
\centering
\includegraphics[width=1.0\textwidth]{./runs/NStep + PER/results.png}
\caption{NStep + PER. Combining the two methods counter the unstable loss function for training in PER.}
\end{figure}
\item [7.] \textbf{Noisy DQN}
\begin{figure}[H]
\centering
\includegraphics[width=1.0\textwidth]{./runs/Noisy DQN/results.png}
\caption{Noisy DQN. Experiment for sigma = 0.017 gets comparable result with normal DQN. Stability issue persist when sigma is too large.}
\end{figure}
\end{enumerate}
\newpage
\item [3.] \textbf{Any other findings}
I implemented Extra credit Noisy DQN. Helpful commands to run in ./commands/4.8.sh Found that when sigma is too large, for example $\sigma=0.5$. The model may not converge to optimal performance. Intuitively, the Noisy linear layer shall improve the robustness of the model. But the effect is not obvious as expected.
\end{enumerate}
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%

View File

@@ -1,154 +0,0 @@
hydra:
run:
dir: ./runs/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
sweep:
dir: ./sweeps/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task: []
job:
name: main
chdir: true
override_dirname: ''
id: ???
num: ???
config_name: config
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\Homeworks\hw2\hw2\cfgs
schema: file
provider: main
- path: ''
schema: structured
provider: schema
output_dir: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\runs\2025-10-11\20-24-05_
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@@ -1,7 +0,0 @@
[2025-10-11 20:24:05,493][py.warnings][WARNING] - C:\Users\wuzhe\anaconda3\envs\drl_hw2\lib\site-packages\torch\cuda\__init__.py:173: UserWarning:
NVIDIA GeForce RTX 5090 with CUDA capability sm_120 is not compatible with the current PyTorch installation.
The current PyTorch install supports CUDA capabilities sm_37 sm_50 sm_60 sm_61 sm_70 sm_75 sm_80 sm_86 compute_37.
If you want to use the NVIDIA GeForce RTX 5090 GPU with PyTorch, please check the instructions at https://pytorch.org/get-started/locally/
warnings.warn(incompatible_device_warn.format(device_name, capability, " ".join(arch_list), device_name))

View File

@@ -1,154 +0,0 @@
hydra:
run:
dir: ./runs/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
sweep:
dir: ./sweeps/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task: []
job:
name: main
chdir: true
override_dirname: ''
id: ???
num: ???
config_name: config
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\Homeworks\hw2\hw2\cfgs
schema: file
provider: main
- path: ''
schema: structured
provider: schema
output_dir: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\runs\2025-10-11\20-31-16_
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@@ -1 +0,0 @@
[]

View File

@@ -1,7 +0,0 @@
[2025-10-11 20:31:16,113][py.warnings][WARNING] - C:\Users\wuzhe\anaconda3\envs\drl_hw2\lib\site-packages\torch\cuda\__init__.py:173: UserWarning:
NVIDIA GeForce RTX 5090 with CUDA capability sm_120 is not compatible with the current PyTorch installation.
The current PyTorch install supports CUDA capabilities sm_37 sm_50 sm_60 sm_61 sm_70 sm_75 sm_80 sm_86 compute_37.
If you want to use the NVIDIA GeForce RTX 5090 GPU with PyTorch, please check the instructions at https://pytorch.org/get-started/locally/
warnings.warn(incompatible_device_warn.format(device_name, capability, " ".join(arch_list), device_name))

View File

@@ -1,154 +0,0 @@
hydra:
run:
dir: ./runs/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
sweep:
dir: ./sweeps/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task: []
job:
name: main
chdir: true
override_dirname: ''
id: ???
num: ???
config_name: config
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\Homeworks\hw2\hw2\cfgs
schema: file
provider: main
- path: ''
schema: structured
provider: schema
output_dir: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\runs\2025-10-11\20-35-30_
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@@ -1 +0,0 @@
[]

View File

@@ -1,7 +0,0 @@
[2025-10-11 20:35:30,859][py.warnings][WARNING] - C:\Users\wuzhe\anaconda3\envs\drl_hw2\lib\site-packages\torch\cuda\__init__.py:173: UserWarning:
NVIDIA GeForce RTX 5090 with CUDA capability sm_120 is not compatible with the current PyTorch installation.
The current PyTorch install supports CUDA capabilities sm_37 sm_50 sm_60 sm_61 sm_70 sm_75 sm_80 sm_86 compute_37.
If you want to use the NVIDIA GeForce RTX 5090 GPU with PyTorch, please check the instructions at https://pytorch.org/get-started/locally/
warnings.warn(incompatible_device_warn.format(device_name, capability, " ".join(arch_list), device_name))

View File

@@ -1,154 +0,0 @@
hydra:
run:
dir: ./runs/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
sweep:
dir: ./sweeps/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task: []
job:
name: main
chdir: true
override_dirname: ''
id: ???
num: ???
config_name: config
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\Homeworks\hw2\hw2\cfgs
schema: file
provider: main
- path: ''
schema: structured
provider: schema
output_dir: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\runs\2025-10-11\20-39-45_
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@@ -1 +0,0 @@
[]

View File

@@ -1,22 +0,0 @@
[2025-10-11 20:39:45,474][py.warnings][WARNING] - C:\Users\wuzhe\anaconda3\envs\drl_hw2\lib\site-packages\torch\cuda\__init__.py:283: UserWarning:
Found GPU0 NVIDIA GeForce RTX 5090 which is of cuda capability 12.0.
Minimum and Maximum cuda capability supported by this version of PyTorch is
(6.1) - (9.0)
warnings.warn(
[2025-10-11 20:39:45,474][py.warnings][WARNING] - C:\Users\wuzhe\anaconda3\envs\drl_hw2\lib\site-packages\torch\cuda\__init__.py:304: UserWarning:
Please install PyTorch with a following CUDA
configurations: 12.8 12.9 following instructions at
https://pytorch.org/get-started/locally/
warnings.warn(matched_cuda_warn.format(matched_arches))
[2025-10-11 20:39:45,476][py.warnings][WARNING] - C:\Users\wuzhe\anaconda3\envs\drl_hw2\lib\site-packages\torch\cuda\__init__.py:326: UserWarning:
NVIDIA GeForce RTX 5090 with CUDA capability sm_120 is not compatible with the current PyTorch installation.
The current PyTorch install supports CUDA capabilities sm_61 sm_70 sm_75 sm_80 sm_86 sm_90.
If you want to use the NVIDIA GeForce RTX 5090 GPU with PyTorch, please check the instructions at https://pytorch.org/get-started/locally/
warnings.warn(
[2025-10-11 20:39:47,115][__main__][INFO] - Training for 50000 timesteps with NormalQNetwork and NormalReplayBuffer

View File

@@ -1,154 +0,0 @@
hydra:
run:
dir: ./runs/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
sweep:
dir: ./sweeps/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task: []
job:
name: main
chdir: true
override_dirname: ''
id: ???
num: ???
config_name: config
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\Homeworks\hw2\hw2\cfgs
schema: file
provider: main
- path: ''
schema: structured
provider: schema
output_dir: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\runs\2025-10-11\20-41-09_
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@@ -1 +0,0 @@
[]

View File

@@ -1,22 +0,0 @@
[2025-10-11 20:41:09,978][py.warnings][WARNING] - C:\Users\wuzhe\anaconda3\envs\drl_hw2\lib\site-packages\torch\cuda\__init__.py:283: UserWarning:
Found GPU0 NVIDIA GeForce RTX 5090 which is of cuda capability 12.0.
Minimum and Maximum cuda capability supported by this version of PyTorch is
(6.1) - (9.0)
warnings.warn(
[2025-10-11 20:41:09,979][py.warnings][WARNING] - C:\Users\wuzhe\anaconda3\envs\drl_hw2\lib\site-packages\torch\cuda\__init__.py:304: UserWarning:
Please install PyTorch with a following CUDA
configurations: 12.8 12.9 following instructions at
https://pytorch.org/get-started/locally/
warnings.warn(matched_cuda_warn.format(matched_arches))
[2025-10-11 20:41:09,979][py.warnings][WARNING] - C:\Users\wuzhe\anaconda3\envs\drl_hw2\lib\site-packages\torch\cuda\__init__.py:326: UserWarning:
NVIDIA GeForce RTX 5090 with CUDA capability sm_120 is not compatible with the current PyTorch installation.
The current PyTorch install supports CUDA capabilities sm_61 sm_70 sm_75 sm_80 sm_86 sm_90.
If you want to use the NVIDIA GeForce RTX 5090 GPU with PyTorch, please check the instructions at https://pytorch.org/get-started/locally/
warnings.warn(
[2025-10-11 20:41:11,670][__main__][INFO] - Training for 50000 timesteps with NormalQNetwork and NormalReplayBuffer

View File

@@ -1,33 +0,0 @@
seed: 42
env_name: CartPole-v1
train:
nstep: ${buffer.nstep}
timesteps: 50000
batch_size: 128
test_every: 2500
eps_max: 1
eps_min: 0.05
eps_steps: 12500
start_steps: 0
plot_interval: 2000
eval_interval: 2000
eval_episodes: 10
agent:
gamma: 0.99
lr: 0.002
tau: 0.1
nstep: ${buffer.nstep}
target_update_interval: 3
hidden_size: 64
activation:
_target_: torch.nn.ELU
use_dueling: false
use_double: false
buffer:
capacity: 50000
use_per: false
nstep: 1
gamma: ${agent.gamma}
per_alpha: 0.7
per_beta: 0.4
per_eps: 0.01

View File

@@ -1,154 +0,0 @@
hydra:
run:
dir: ./runs/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
sweep:
dir: ./sweeps/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task: []
job:
name: main
chdir: true
override_dirname: ''
id: ???
num: ???
config_name: config
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\Homeworks\hw2\hw2\cfgs
schema: file
provider: main
- path: ''
schema: structured
provider: schema
output_dir: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\runs\2025-10-11\20-42-26_
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@@ -1 +0,0 @@
[]

View File

@@ -1,22 +0,0 @@
[2025-10-11 20:42:26,843][py.warnings][WARNING] - C:\Users\wuzhe\anaconda3\envs\drl_hw2\lib\site-packages\torch\cuda\__init__.py:283: UserWarning:
Found GPU0 NVIDIA GeForce RTX 5090 which is of cuda capability 12.0.
Minimum and Maximum cuda capability supported by this version of PyTorch is
(6.1) - (9.0)
warnings.warn(
[2025-10-11 20:42:26,844][py.warnings][WARNING] - C:\Users\wuzhe\anaconda3\envs\drl_hw2\lib\site-packages\torch\cuda\__init__.py:304: UserWarning:
Please install PyTorch with a following CUDA
configurations: 12.8 12.9 following instructions at
https://pytorch.org/get-started/locally/
warnings.warn(matched_cuda_warn.format(matched_arches))
[2025-10-11 20:42:26,846][py.warnings][WARNING] - C:\Users\wuzhe\anaconda3\envs\drl_hw2\lib\site-packages\torch\cuda\__init__.py:326: UserWarning:
NVIDIA GeForce RTX 5090 with CUDA capability sm_120 is not compatible with the current PyTorch installation.
The current PyTorch install supports CUDA capabilities sm_61 sm_70 sm_75 sm_80 sm_86 sm_90.
If you want to use the NVIDIA GeForce RTX 5090 GPU with PyTorch, please check the instructions at https://pytorch.org/get-started/locally/
warnings.warn(
[2025-10-11 20:42:28,580][__main__][INFO] - Training for 50000 timesteps with NormalQNetwork and NormalReplayBuffer

View File

@@ -1,33 +0,0 @@
seed: 42
env_name: CartPole-v1
train:
nstep: ${buffer.nstep}
timesteps: 50000
batch_size: 128
test_every: 2500
eps_max: 1
eps_min: 0.05
eps_steps: 12500
start_steps: 0
plot_interval: 2000
eval_interval: 2000
eval_episodes: 10
agent:
gamma: 0.99
lr: 0.002
tau: 0.1
nstep: ${buffer.nstep}
target_update_interval: 3
hidden_size: 64
activation:
_target_: torch.nn.ELU
use_dueling: false
use_double: false
buffer:
capacity: 50000
use_per: false
nstep: 1
gamma: ${agent.gamma}
per_alpha: 0.7
per_beta: 0.4
per_eps: 0.01

View File

@@ -1,154 +0,0 @@
hydra:
run:
dir: ./runs/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
sweep:
dir: ./sweeps/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task: []
job:
name: main
chdir: true
override_dirname: ''
id: ???
num: ???
config_name: config
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\Homeworks\hw2\hw2\cfgs
schema: file
provider: main
- path: ''
schema: structured
provider: schema
output_dir: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\runs\2025-10-11\20-45-12_
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@@ -1 +0,0 @@
[]

View File

@@ -1,22 +0,0 @@
[2025-10-11 20:45:12,694][py.warnings][WARNING] - C:\Users\wuzhe\anaconda3\envs\drl_hw2\lib\site-packages\torch\cuda\__init__.py:283: UserWarning:
Found GPU0 NVIDIA GeForce RTX 5090 which is of cuda capability 12.0.
Minimum and Maximum cuda capability supported by this version of PyTorch is
(6.1) - (9.0)
warnings.warn(
[2025-10-11 20:45:12,694][py.warnings][WARNING] - C:\Users\wuzhe\anaconda3\envs\drl_hw2\lib\site-packages\torch\cuda\__init__.py:304: UserWarning:
Please install PyTorch with a following CUDA
configurations: 12.8 12.9 following instructions at
https://pytorch.org/get-started/locally/
warnings.warn(matched_cuda_warn.format(matched_arches))
[2025-10-11 20:45:12,696][py.warnings][WARNING] - C:\Users\wuzhe\anaconda3\envs\drl_hw2\lib\site-packages\torch\cuda\__init__.py:326: UserWarning:
NVIDIA GeForce RTX 5090 with CUDA capability sm_120 is not compatible with the current PyTorch installation.
The current PyTorch install supports CUDA capabilities sm_61 sm_70 sm_75 sm_80 sm_86 sm_90.
If you want to use the NVIDIA GeForce RTX 5090 GPU with PyTorch, please check the instructions at https://pytorch.org/get-started/locally/
warnings.warn(
[2025-10-11 20:45:14,422][__main__][INFO] - Training for 50000 timesteps with NormalQNetwork and NormalReplayBuffer

View File

@@ -1,33 +0,0 @@
seed: 42
env_name: CartPole-v1
train:
nstep: ${buffer.nstep}
timesteps: 50000
batch_size: 128
test_every: 2500
eps_max: 1
eps_min: 0.05
eps_steps: 12500
start_steps: 0
plot_interval: 2000
eval_interval: 2000
eval_episodes: 10
agent:
gamma: 0.99
lr: 0.002
tau: 0.1
nstep: ${buffer.nstep}
target_update_interval: 3
hidden_size: 64
activation:
_target_: torch.nn.ELU
use_dueling: false
use_double: false
buffer:
capacity: 50000
use_per: false
nstep: 1
gamma: ${agent.gamma}
per_alpha: 0.7
per_beta: 0.4
per_eps: 0.01

View File

@@ -1,154 +0,0 @@
hydra:
run:
dir: ./runs/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
sweep:
dir: ./sweeps/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task: []
job:
name: main
chdir: true
override_dirname: ''
id: ???
num: ???
config_name: config
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\Homeworks\hw2\hw2\cfgs
schema: file
provider: main
- path: ''
schema: structured
provider: schema
output_dir: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\runs\2025-10-11\20-45-23_
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@@ -1 +0,0 @@
[]

View File

@@ -1,22 +0,0 @@
[2025-10-11 20:45:23,927][py.warnings][WARNING] - C:\Users\wuzhe\anaconda3\envs\drl_hw2\lib\site-packages\torch\cuda\__init__.py:283: UserWarning:
Found GPU0 NVIDIA GeForce RTX 5090 which is of cuda capability 12.0.
Minimum and Maximum cuda capability supported by this version of PyTorch is
(6.1) - (9.0)
warnings.warn(
[2025-10-11 20:45:23,928][py.warnings][WARNING] - C:\Users\wuzhe\anaconda3\envs\drl_hw2\lib\site-packages\torch\cuda\__init__.py:304: UserWarning:
Please install PyTorch with a following CUDA
configurations: 12.8 12.9 following instructions at
https://pytorch.org/get-started/locally/
warnings.warn(matched_cuda_warn.format(matched_arches))
[2025-10-11 20:45:23,930][py.warnings][WARNING] - C:\Users\wuzhe\anaconda3\envs\drl_hw2\lib\site-packages\torch\cuda\__init__.py:326: UserWarning:
NVIDIA GeForce RTX 5090 with CUDA capability sm_120 is not compatible with the current PyTorch installation.
The current PyTorch install supports CUDA capabilities sm_61 sm_70 sm_75 sm_80 sm_86 sm_90.
If you want to use the NVIDIA GeForce RTX 5090 GPU with PyTorch, please check the instructions at https://pytorch.org/get-started/locally/
warnings.warn(
[2025-10-11 20:45:25,714][__main__][INFO] - Training for 50000 timesteps with NormalQNetwork and NormalReplayBuffer

View File

@@ -1,33 +0,0 @@
seed: 42
env_name: CartPole-v1
train:
nstep: ${buffer.nstep}
timesteps: 50000
batch_size: 128
test_every: 2500
eps_max: 1
eps_min: 0.05
eps_steps: 12500
start_steps: 0
plot_interval: 2000
eval_interval: 2000
eval_episodes: 10
agent:
gamma: 0.99
lr: 0.002
tau: 0.1
nstep: ${buffer.nstep}
target_update_interval: 3
hidden_size: 64
activation:
_target_: torch.nn.ELU
use_dueling: false
use_double: false
buffer:
capacity: 50000
use_per: false
nstep: 1
gamma: ${agent.gamma}
per_alpha: 0.7
per_beta: 0.4
per_eps: 0.01

View File

@@ -1,154 +0,0 @@
hydra:
run:
dir: ./runs/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
sweep:
dir: ./sweeps/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task: []
job:
name: main
chdir: true
override_dirname: ''
id: ???
num: ???
config_name: config
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\Homeworks\hw2\hw2\cfgs
schema: file
provider: main
- path: ''
schema: structured
provider: schema
output_dir: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\runs\2025-10-11\20-55-32_
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@@ -1 +0,0 @@
[]

View File

@@ -1,22 +0,0 @@
[2025-10-11 20:55:32,238][py.warnings][WARNING] - C:\Users\wuzhe\anaconda3\envs\drl_hw2\lib\site-packages\torch\cuda\__init__.py:283: UserWarning:
Found GPU0 NVIDIA GeForce RTX 5090 which is of cuda capability 12.0.
Minimum and Maximum cuda capability supported by this version of PyTorch is
(6.1) - (9.0)
warnings.warn(
[2025-10-11 20:55:32,238][py.warnings][WARNING] - C:\Users\wuzhe\anaconda3\envs\drl_hw2\lib\site-packages\torch\cuda\__init__.py:304: UserWarning:
Please install PyTorch with a following CUDA
configurations: 12.8 12.9 following instructions at
https://pytorch.org/get-started/locally/
warnings.warn(matched_cuda_warn.format(matched_arches))
[2025-10-11 20:55:32,240][py.warnings][WARNING] - C:\Users\wuzhe\anaconda3\envs\drl_hw2\lib\site-packages\torch\cuda\__init__.py:326: UserWarning:
NVIDIA GeForce RTX 5090 with CUDA capability sm_120 is not compatible with the current PyTorch installation.
The current PyTorch install supports CUDA capabilities sm_61 sm_70 sm_75 sm_80 sm_86 sm_90.
If you want to use the NVIDIA GeForce RTX 5090 GPU with PyTorch, please check the instructions at https://pytorch.org/get-started/locally/
warnings.warn(
[2025-10-11 20:55:33,876][__main__][INFO] - Training for 50000 timesteps with NormalQNetwork and NormalReplayBuffer

View File

@@ -1,33 +0,0 @@
seed: 42
env_name: CartPole-v1
train:
nstep: ${buffer.nstep}
timesteps: 50000
batch_size: 128
test_every: 2500
eps_max: 1
eps_min: 0.05
eps_steps: 12500
start_steps: 0
plot_interval: 2000
eval_interval: 2000
eval_episodes: 10
agent:
gamma: 0.99
lr: 0.002
tau: 0.1
nstep: ${buffer.nstep}
target_update_interval: 3
hidden_size: 64
activation:
_target_: torch.nn.ELU
use_dueling: false
use_double: false
buffer:
capacity: 50000
use_per: false
nstep: 1
gamma: ${agent.gamma}
per_alpha: 0.7
per_beta: 0.4
per_eps: 0.01

View File

@@ -1,154 +0,0 @@
hydra:
run:
dir: ./runs/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
sweep:
dir: ./sweeps/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task: []
job:
name: main
chdir: true
override_dirname: ''
id: ???
num: ???
config_name: config
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\Homeworks\hw2\hw2\cfgs
schema: file
provider: main
- path: ''
schema: structured
provider: schema
output_dir: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\runs\2025-10-11\20-59-59_
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@@ -1 +0,0 @@
[]

View File

@@ -1 +0,0 @@
[2025-10-11 21:00:01,190][__main__][INFO] - Training for 50000 timesteps with NormalQNetwork and NormalReplayBuffer

View File

@@ -1,33 +0,0 @@
seed: 42
env_name: CartPole-v1
train:
nstep: ${buffer.nstep}
timesteps: 50000
batch_size: 128
test_every: 2500
eps_max: 1
eps_min: 0.05
eps_steps: 12500
start_steps: 0
plot_interval: 2000
eval_interval: 2000
eval_episodes: 10
agent:
gamma: 0.99
lr: 0.002
tau: 0.1
nstep: ${buffer.nstep}
target_update_interval: 3
hidden_size: 64
activation:
_target_: torch.nn.ELU
use_dueling: false
use_double: false
buffer:
capacity: 50000
use_per: false
nstep: 1
gamma: ${agent.gamma}
per_alpha: 0.7
per_beta: 0.4
per_eps: 0.01

View File

@@ -1,154 +0,0 @@
hydra:
run:
dir: ./runs/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
sweep:
dir: ./sweeps/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task: []
job:
name: main
chdir: true
override_dirname: ''
id: ???
num: ???
config_name: config
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\Homeworks\hw2\hw2\cfgs
schema: file
provider: main
- path: ''
schema: structured
provider: schema
output_dir: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\runs\2025-10-11\21-00-50_
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@@ -1 +0,0 @@
[]

View File

@@ -1 +0,0 @@
[2025-10-11 21:00:52,388][__main__][INFO] - Training for 50000 timesteps with NormalQNetwork and NormalReplayBuffer

View File

@@ -1,33 +0,0 @@
seed: 42
env_name: CartPole-v1
train:
nstep: ${buffer.nstep}
timesteps: 50000
batch_size: 128
test_every: 2500
eps_max: 1
eps_min: 0.05
eps_steps: 12500
start_steps: 0
plot_interval: 2000
eval_interval: 2000
eval_episodes: 10
agent:
gamma: 0.99
lr: 0.002
tau: 0.1
nstep: ${buffer.nstep}
target_update_interval: 3
hidden_size: 64
activation:
_target_: torch.nn.ELU
use_dueling: false
use_double: false
buffer:
capacity: 50000
use_per: false
nstep: 1
gamma: ${agent.gamma}
per_alpha: 0.7
per_beta: 0.4
per_eps: 0.01

View File

@@ -1,154 +0,0 @@
hydra:
run:
dir: ./runs/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
sweep:
dir: ./sweeps/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task: []
job:
name: main
chdir: true
override_dirname: ''
id: ???
num: ???
config_name: config
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\Homeworks\hw2\hw2\cfgs
schema: file
provider: main
- path: ''
schema: structured
provider: schema
output_dir: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\runs\2025-10-11\21-03-24_
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@@ -1 +0,0 @@
[]

View File

@@ -1 +0,0 @@
[2025-10-11 21:03:26,154][__main__][INFO] - Training for 50000 timesteps with NormalQNetwork and NormalReplayBuffer

View File

@@ -1,33 +0,0 @@
seed: 42
env_name: CartPole-v1
train:
nstep: ${buffer.nstep}
timesteps: 50000
batch_size: 128
test_every: 2500
eps_max: 1
eps_min: 0.05
eps_steps: 12500
start_steps: 0
plot_interval: 2000
eval_interval: 2000
eval_episodes: 10
agent:
gamma: 0.99
lr: 0.002
tau: 0.1
nstep: ${buffer.nstep}
target_update_interval: 3
hidden_size: 64
activation:
_target_: torch.nn.ELU
use_dueling: false
use_double: false
buffer:
capacity: 50000
use_per: false
nstep: 1
gamma: ${agent.gamma}
per_alpha: 0.7
per_beta: 0.4
per_eps: 0.01

View File

@@ -1,154 +0,0 @@
hydra:
run:
dir: ./runs/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
sweep:
dir: ./sweeps/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task: []
job:
name: main
chdir: true
override_dirname: ''
id: ???
num: ???
config_name: config
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\Homeworks\hw2\hw2\cfgs
schema: file
provider: main
- path: ''
schema: structured
provider: schema
output_dir: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\runs\2025-10-11\21-03-35_
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@@ -1 +0,0 @@
[]

View File

@@ -1 +0,0 @@
[2025-10-11 21:03:36,838][__main__][INFO] - Training for 50000 timesteps with NormalQNetwork and NormalReplayBuffer

View File

@@ -1,33 +0,0 @@
seed: 42
env_name: CartPole-v1
train:
nstep: ${buffer.nstep}
timesteps: 50000
batch_size: 128
test_every: 2500
eps_max: 1
eps_min: 0.05
eps_steps: 12500
start_steps: 0
plot_interval: 2000
eval_interval: 2000
eval_episodes: 10
agent:
gamma: 0.99
lr: 0.002
tau: 0.1
nstep: ${buffer.nstep}
target_update_interval: 3
hidden_size: 64
activation:
_target_: torch.nn.ELU
use_dueling: false
use_double: false
buffer:
capacity: 50000
use_per: false
nstep: 1
gamma: ${agent.gamma}
per_alpha: 0.7
per_beta: 0.4
per_eps: 0.01

View File

@@ -1,154 +0,0 @@
hydra:
run:
dir: ./runs/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
sweep:
dir: ./sweeps/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task: []
job:
name: main
chdir: true
override_dirname: ''
id: ???
num: ???
config_name: config
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\Homeworks\hw2\hw2\cfgs
schema: file
provider: main
- path: ''
schema: structured
provider: schema
output_dir: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\runs\2025-10-11\21-05-11_
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@@ -1 +0,0 @@
[]

View File

@@ -1 +0,0 @@
[2025-10-11 21:05:12,880][__main__][INFO] - Training for 50000 timesteps with NormalQNetwork and NormalReplayBuffer

View File

@@ -1,33 +0,0 @@
seed: 42
env_name: CartPole-v1
train:
nstep: ${buffer.nstep}
timesteps: 50000
batch_size: 128
test_every: 2500
eps_max: 1
eps_min: 0.05
eps_steps: 12500
start_steps: 0
plot_interval: 2000
eval_interval: 2000
eval_episodes: 10
agent:
gamma: 0.99
lr: 0.002
tau: 0.1
nstep: ${buffer.nstep}
target_update_interval: 3
hidden_size: 64
activation:
_target_: torch.nn.ELU
use_dueling: false
use_double: false
buffer:
capacity: 50000
use_per: false
nstep: 1
gamma: ${agent.gamma}
per_alpha: 0.7
per_beta: 0.4
per_eps: 0.01

View File

@@ -1,154 +0,0 @@
hydra:
run:
dir: ./runs/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
sweep:
dir: ./sweeps/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task: []
job:
name: main
chdir: true
override_dirname: ''
id: ???
num: ???
config_name: config
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\Homeworks\hw2\hw2\cfgs
schema: file
provider: main
- path: ''
schema: structured
provider: schema
output_dir: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\runs\2025-10-11\21-07-21_
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@@ -1 +0,0 @@
[]

View File

@@ -1 +0,0 @@
[2025-10-11 21:07:22,911][__main__][INFO] - Training for 50000 timesteps with NormalQNetwork and NormalReplayBuffer

View File

@@ -1,33 +0,0 @@
seed: 42
env_name: CartPole-v1
train:
nstep: ${buffer.nstep}
timesteps: 50000
batch_size: 128
test_every: 2500
eps_max: 1
eps_min: 0.05
eps_steps: 12500
start_steps: 0
plot_interval: 2000
eval_interval: 2000
eval_episodes: 10
agent:
gamma: 0.99
lr: 0.002
tau: 0.1
nstep: ${buffer.nstep}
target_update_interval: 3
hidden_size: 64
activation:
_target_: torch.nn.ELU
use_dueling: false
use_double: false
buffer:
capacity: 50000
use_per: false
nstep: 1
gamma: ${agent.gamma}
per_alpha: 0.7
per_beta: 0.4
per_eps: 0.01

View File

@@ -1,154 +0,0 @@
hydra:
run:
dir: ./runs/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
sweep:
dir: ./sweeps/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task: []
job:
name: main
chdir: true
override_dirname: ''
id: ???
num: ???
config_name: config
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\Homeworks\hw2\hw2\cfgs
schema: file
provider: main
- path: ''
schema: structured
provider: schema
output_dir: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\runs\2025-10-11\21-07-44_
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@@ -1 +0,0 @@
[]

View File

@@ -1 +0,0 @@
[2025-10-11 21:07:45,823][__main__][INFO] - Training for 50000 timesteps with NormalQNetwork and NormalReplayBuffer

View File

@@ -1,33 +0,0 @@
seed: 42
env_name: CartPole-v1
train:
nstep: ${buffer.nstep}
timesteps: 50000
batch_size: 128
test_every: 2500
eps_max: 1
eps_min: 0.05
eps_steps: 12500
start_steps: 0
plot_interval: 2000
eval_interval: 2000
eval_episodes: 10
agent:
gamma: 0.99
lr: 0.002
tau: 0.1
nstep: ${buffer.nstep}
target_update_interval: 3
hidden_size: 64
activation:
_target_: torch.nn.ELU
use_dueling: false
use_double: false
buffer:
capacity: 50000
use_per: false
nstep: 1
gamma: ${agent.gamma}
per_alpha: 0.7
per_beta: 0.4
per_eps: 0.01

View File

@@ -1,154 +0,0 @@
hydra:
run:
dir: ./runs/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
sweep:
dir: ./sweeps/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task: []
job:
name: main
chdir: true
override_dirname: ''
id: ???
num: ???
config_name: config
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\Homeworks\hw2\hw2\cfgs
schema: file
provider: main
- path: ''
schema: structured
provider: schema
output_dir: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\runs\2025-10-11\21-08-54_
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@@ -1 +0,0 @@
[]

View File

@@ -1 +0,0 @@
[2025-10-11 21:08:56,669][__main__][INFO] - Training for 50000 timesteps with NormalQNetwork and NormalReplayBuffer

View File

@@ -1,33 +0,0 @@
seed: 42
env_name: CartPole-v1
train:
nstep: ${buffer.nstep}
timesteps: 50000
batch_size: 128
test_every: 2500
eps_max: 1
eps_min: 0.05
eps_steps: 12500
start_steps: 0
plot_interval: 2000
eval_interval: 2000
eval_episodes: 10
agent:
gamma: 0.99
lr: 0.002
tau: 0.1
nstep: ${buffer.nstep}
target_update_interval: 3
hidden_size: 64
activation:
_target_: torch.nn.ELU
use_dueling: false
use_double: false
buffer:
capacity: 50000
use_per: false
nstep: 1
gamma: ${agent.gamma}
per_alpha: 0.7
per_beta: 0.4
per_eps: 0.01

View File

@@ -1,154 +0,0 @@
hydra:
run:
dir: ./runs/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
sweep:
dir: ./sweeps/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task: []
job:
name: main
chdir: true
override_dirname: ''
id: ???
num: ???
config_name: config
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\Homeworks\hw2\hw2\cfgs
schema: file
provider: main
- path: ''
schema: structured
provider: schema
output_dir: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\runs\2025-10-11\21-09-36_
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@@ -1 +0,0 @@
[]

View File

@@ -1 +0,0 @@
[2025-10-11 21:09:38,404][__main__][INFO] - Training for 50000 timesteps with NormalQNetwork and NormalReplayBuffer

View File

@@ -1,33 +0,0 @@
seed: 42
env_name: CartPole-v1
train:
nstep: ${buffer.nstep}
timesteps: 50000
batch_size: 128
test_every: 2500
eps_max: 1
eps_min: 0.05
eps_steps: 12500
start_steps: 0
plot_interval: 2000
eval_interval: 2000
eval_episodes: 10
agent:
gamma: 0.99
lr: 0.002
tau: 0.1
nstep: ${buffer.nstep}
target_update_interval: 3
hidden_size: 64
activation:
_target_: torch.nn.ELU
use_dueling: false
use_double: false
buffer:
capacity: 50000
use_per: false
nstep: 1
gamma: ${agent.gamma}
per_alpha: 0.7
per_beta: 0.4
per_eps: 0.01

View File

@@ -1,154 +0,0 @@
hydra:
run:
dir: ./runs/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
sweep:
dir: ./sweeps/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task: []
job:
name: main
chdir: true
override_dirname: ''
id: ???
num: ???
config_name: config
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\Homeworks\hw2\hw2\cfgs
schema: file
provider: main
- path: ''
schema: structured
provider: schema
output_dir: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\runs\2025-10-11\21-10-00_
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@@ -1 +0,0 @@
[]

View File

@@ -1 +0,0 @@
[2025-10-11 21:10:02,340][__main__][INFO] - Training for 50000 timesteps with NormalQNetwork and NormalReplayBuffer

View File

@@ -1,33 +0,0 @@
seed: 42
env_name: CartPole-v1
train:
nstep: ${buffer.nstep}
timesteps: 50000
batch_size: 128
test_every: 2500
eps_max: 1
eps_min: 0.05
eps_steps: 12500
start_steps: 0
plot_interval: 2000
eval_interval: 2000
eval_episodes: 10
agent:
gamma: 0.99
lr: 0.002
tau: 0.1
nstep: ${buffer.nstep}
target_update_interval: 3
hidden_size: 64
activation:
_target_: torch.nn.ELU
use_dueling: false
use_double: false
buffer:
capacity: 50000
use_per: false
nstep: 1
gamma: ${agent.gamma}
per_alpha: 0.7
per_beta: 0.4
per_eps: 0.01

View File

@@ -1,154 +0,0 @@
hydra:
run:
dir: ./runs/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
sweep:
dir: ./sweeps/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task: []
job:
name: main
chdir: true
override_dirname: ''
id: ???
num: ???
config_name: config
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\Homeworks\hw2\hw2\cfgs
schema: file
provider: main
- path: ''
schema: structured
provider: schema
output_dir: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\runs\2025-10-11\21-15-36_
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@@ -1 +0,0 @@
[]

View File

@@ -1 +0,0 @@
[2025-10-11 21:15:37,961][__main__][INFO] - Training for 50000 timesteps with NormalQNetwork and NormalReplayBuffer

View File

@@ -1,33 +0,0 @@
seed: 42
env_name: CartPole-v1
train:
nstep: ${buffer.nstep}
timesteps: 50000
batch_size: 128
test_every: 2500
eps_max: 1
eps_min: 0.05
eps_steps: 12500
start_steps: 0
plot_interval: 2000
eval_interval: 2000
eval_episodes: 10
agent:
gamma: 0.99
lr: 0.002
tau: 0.1
nstep: ${buffer.nstep}
target_update_interval: 3
hidden_size: 64
activation:
_target_: torch.nn.ELU
use_dueling: false
use_double: false
buffer:
capacity: 50000
use_per: false
nstep: 1
gamma: ${agent.gamma}
per_alpha: 0.7
per_beta: 0.4
per_eps: 0.01

View File

@@ -1,154 +0,0 @@
hydra:
run:
dir: ./runs/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
sweep:
dir: ./sweeps/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task: []
job:
name: main
chdir: true
override_dirname: ''
id: ???
num: ???
config_name: config
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\Homeworks\hw2\hw2\cfgs
schema: file
provider: main
- path: ''
schema: structured
provider: schema
output_dir: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\runs\2025-10-11\21-16-27_
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@@ -1 +0,0 @@
[]

View File

@@ -1 +0,0 @@
[2025-10-11 21:16:28,918][__main__][INFO] - Training for 50000 timesteps with NormalQNetwork and NormalReplayBuffer

View File

@@ -1,33 +0,0 @@
seed: 42
env_name: CartPole-v1
train:
nstep: ${buffer.nstep}
timesteps: 50000
batch_size: 128
test_every: 2500
eps_max: 1
eps_min: 0.05
eps_steps: 12500
start_steps: 0
plot_interval: 2000
eval_interval: 2000
eval_episodes: 10
agent:
gamma: 0.99
lr: 0.002
tau: 0.1
nstep: ${buffer.nstep}
target_update_interval: 3
hidden_size: 64
activation:
_target_: torch.nn.ELU
use_dueling: false
use_double: false
buffer:
capacity: 50000
use_per: false
nstep: 1
gamma: ${agent.gamma}
per_alpha: 0.7
per_beta: 0.4
per_eps: 0.01

View File

@@ -1,154 +0,0 @@
hydra:
run:
dir: ./runs/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
sweep:
dir: ./sweeps/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task: []
job:
name: main
chdir: true
override_dirname: ''
id: ???
num: ???
config_name: config
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\Homeworks\hw2\hw2\cfgs
schema: file
provider: main
- path: ''
schema: structured
provider: schema
output_dir: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\runs\2025-10-11\21-16-42_
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@@ -1 +0,0 @@
[]

View File

@@ -1 +0,0 @@
[2025-10-11 21:16:44,069][__main__][INFO] - Training for 50000 timesteps with NormalQNetwork and NormalReplayBuffer

View File

@@ -1,33 +0,0 @@
seed: 42
env_name: CartPole-v1
train:
nstep: ${buffer.nstep}
timesteps: 50000
batch_size: 128
test_every: 2500
eps_max: 1
eps_min: 0.05
eps_steps: 12500
start_steps: 0
plot_interval: 2000
eval_interval: 2000
eval_episodes: 10
agent:
gamma: 0.99
lr: 0.002
tau: 0.1
nstep: ${buffer.nstep}
target_update_interval: 3
hidden_size: 64
activation:
_target_: torch.nn.ELU
use_dueling: false
use_double: false
buffer:
capacity: 50000
use_per: false
nstep: 1
gamma: ${agent.gamma}
per_alpha: 0.7
per_beta: 0.4
per_eps: 0.01

View File

@@ -1,154 +0,0 @@
hydra:
run:
dir: ./runs/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
sweep:
dir: ./sweeps/${now:%Y-%m-%d}/${now:%H-%M-%S}_${hydra.job.override_dirname}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task: []
job:
name: main
chdir: true
override_dirname: ''
id: ???
num: ???
config_name: config
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\Homeworks\hw2\hw2\cfgs
schema: file
provider: main
- path: ''
schema: structured
provider: schema
output_dir: D:\Documents\Nextcloud\Documents\Project WUSTL\Academic\2025_Fall\CSE5100\runs\2025-10-11\21-17-17_
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@@ -1 +0,0 @@
[]

View File

@@ -1 +0,0 @@
[2025-10-11 21:17:19,615][__main__][INFO] - Training for 50000 timesteps with NormalQNetwork and NormalReplayBuffer

View File

@@ -1,33 +0,0 @@
seed: 42
env_name: CartPole-v1
train:
nstep: ${buffer.nstep}
timesteps: 50000
batch_size: 128
test_every: 2500
eps_max: 1
eps_min: 0.05
eps_steps: 12500
start_steps: 0
plot_interval: 2000
eval_interval: 2000
eval_episodes: 10
agent:
gamma: 0.99
lr: 0.002
tau: 0.1
nstep: ${buffer.nstep}
target_update_interval: 3
hidden_size: 64
activation:
_target_: torch.nn.ELU
use_dueling: false
use_double: false
buffer:
capacity: 50000
use_per: false
nstep: 1
gamma: ${agent.gamma}
per_alpha: 0.7
per_beta: 0.4
per_eps: 0.01

Some files were not shown because too many files have changed in this diff Show More