in jat/eval/rl/wrappers.py [0:0]
def reset(self, **kwargs) -> Tuple[np.ndarray, Dict]:
self.env.reset(**kwargs)
noops = self.unwrapped.np_random.integers(1, self.noop_max + 1)
for _ in range(noops):
observation, reward, terminated, truncated, info = self.env.step(self.noop_action)
if terminated | truncated:
observation, info = self.env.reset(**kwargs)
return observation, info