HyperAI
HyperAI超神经
首页
算力平台
文档
资讯
论文
教程
数据集
百科
SOTA
LLM 模型天梯
GPU 天梯
顶会
开源项目
全站搜索
关于
服务条款
隐私政策
中文
HyperAI
HyperAI
Toggle Sidebar
全站搜索…
⌘
K
Command Palette
Search for a command to run...
Console
Sign In
首页
SOTA
OpenAI Gym
Openai Gym On Hopper V4
Openai Gym On Hopper V4
Metrics
Average Return
Results
Performance results of various models on this benchmark
Columns
Model Name
Average Return
Paper Title
Code
MEow
3332.99
Maximum Entropy Reinforcement Learning via Energy-Based Normalizing Flow
TD3
3319.98
Addressing Function Approximation Error in Actor-Critic Methods
SAC
2882.56
Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor
DDPG
1290.24
Continuous control with deep reinforcement learning
PPO
790.77
Proximal Policy Optimization Algorithms
0 of 5 row(s) selected.
Previous
Next
HyperAI
HyperAI超神经
首页
算力平台
文档
资讯
论文
教程
数据集
百科
SOTA
LLM 模型天梯
GPU 天梯
顶会
开源项目
全站搜索
关于
服务条款
隐私政策
中文
HyperAI
HyperAI
Toggle Sidebar
全站搜索…
⌘
K
Command Palette
Search for a command to run...
Console
Sign In
首页
SOTA
OpenAI Gym
Openai Gym On Hopper V4
Openai Gym On Hopper V4
Metrics
Average Return
Results
Performance results of various models on this benchmark
Columns
Model Name
Average Return
Paper Title
Code
MEow
3332.99
Maximum Entropy Reinforcement Learning via Energy-Based Normalizing Flow
TD3
3319.98
Addressing Function Approximation Error in Actor-Critic Methods
SAC
2882.56
Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor
DDPG
1290.24
Continuous control with deep reinforcement learning
PPO
790.77
Proximal Policy Optimization Algorithms
0 of 5 row(s) selected.
Previous
Next
Console
Console