offline-rl
CommunityMaster offline RL with data-driven conservatism.
Authortachyon-beep
Version1.0.0
Installs0
System Documentation
What problem does it solve?
This skill enables practitioners to perform offline reinforcement learning from fixed datasets without interacting with an environment, addressing issues like distribution shift and extrapolation errors by adopting conservative value estimation and robust evaluation strategies.
Core Features & Use Cases
- Conservative offline RL methods: CQL, IQL, and BCQ to mitigate overestimation when data is limited or distribution shifts occur.
- Offline training & evaluation: Train purely from logged data and evaluate using offline techniques to estimate policy performance without live rollout.
- Real-world scenarios: Robotics from logged demonstrations, medical treatment policy learning from historical data, and recommendations improved from historical user interactions.
Quick Start
Prepare a fixed dataset, choose a method (CQL, IQL, or BCQ), train the model, and perform offline evaluation using RIS or model-based estimates.
Dependency Matrix
Required Modules
None requiredComponents
Standard package💻 Claude Code Installation
Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.
Please help me install this Skill: Name: offline-rl Download link: https://github.com/tachyon-beep/hamlet/archive/main.zip#offline-rl Please download this .zip file, extract it, and install it in the .claude/skills/ directory.
Agent Skills Search Helper
Install a tiny helper to your Agent, search and equip skill from 223,000+ vetted skills library on demand.