offline-rl

Community

Master offline RL with data-driven conservatism.

Authortachyon-beep
Version1.0.0
Installs0

System Documentation

What problem does it solve?

This skill enables practitioners to perform offline reinforcement learning from fixed datasets without interacting with an environment, addressing issues like distribution shift and extrapolation errors by adopting conservative value estimation and robust evaluation strategies.

Core Features & Use Cases

  • Conservative offline RL methods: CQL, IQL, and BCQ to mitigate overestimation when data is limited or distribution shifts occur.
  • Offline training & evaluation: Train purely from logged data and evaluate using offline techniques to estimate policy performance without live rollout.
  • Real-world scenarios: Robotics from logged demonstrations, medical treatment policy learning from historical data, and recommendations improved from historical user interactions.

Quick Start

Prepare a fixed dataset, choose a method (CQL, IQL, or BCQ), train the model, and perform offline evaluation using RIS or model-based estimates.

Dependency Matrix

Required Modules

None required

Components

Standard package

💻 Claude Code Installation

Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.

Please help me install this Skill:
Name: offline-rl
Download link: https://github.com/tachyon-beep/hamlet/archive/main.zip#offline-rl

Please download this .zip file, extract it, and install it in the .claude/skills/ directory.
View Source Repository

Agent Skills Search Helper

Install a tiny helper to your Agent, search and equip skill from 223,000+ vetted skills library on demand.