Description
In my setup, I want to:
However, the current ART stack makes this very difficult because:
- RULER scoring (
ruler_score_group and related helpers) rely on Litellm in a way that expects OpenAI-style models.
init_chat_model also wraps everything in a ChatOpenAI instance (see separate issue).
- This means I cannot simply pass
ChatOllama or ChatNVIDIA (LangChain chat models) as the inference/judge model for training.
Practically:
What I’d like
-
A more provider-agnostic design for:
- RULER scoring
- Training
init_chat_model
-
The ability to cleanly use:
ChatOllama (LangChain)
ChatNVIDIA
- or other LangChain
BaseChatModel implementations
-
Without having to hack around Litellm / OpenAI assumptions.
Why this matters
-
ART is otherwise a great framework for agent RL.
-
Many users want to move to:
- Local models (Ollama)
- Different clouds (NVIDIA, etc.)
-
Tight coupling to OpenAI via Litellm in the RULER path makes this significantly harder.
Request
-
Please consider:
- Abstracting RULER to accept any LangChain-compatible
ChatModel for structured scoring.
- Or providing a documented way to plug in non-OpenAI judgment models (e.g. a “judge_fn” that uses arbitrary models).