From 311ddfff4610fb547fd8391880687ec4d6088bb1 Mon Sep 17 00:00:00 2001 From: chenyushuo <297086016@qq.com> Date: Tue, 20 Jan 2026 14:13:20 +0800 Subject: [PATCH] Fix in learn to ask (#107) --- tuner/learn_to_ask/README.md | 2 ++ tuner/learn_to_ask/main.py | 2 -- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/tuner/learn_to_ask/README.md b/tuner/learn_to_ask/README.md index b53f384..b666585 100644 --- a/tuner/learn_to_ask/README.md +++ b/tuner/learn_to_ask/README.md @@ -289,6 +289,8 @@ Also, make sure to update the `model_path` in `tuner/learn_to_ask/main.py` to po > 🔗 Learn more about Tinker Backend: [Tinker Backend Documentation](https://agentscope-ai.github.io/Trinity-RFT/en/main/tutorial/example_tinker_backend.html) +> In this provided example, training is configured for 4 epochs. When using Tinker, the total token consumption is approximately 112 million tokens, resulting in an estimated cost of approximately 18 USD. + ### Launch Training ```bash python tuner/learn_to_ask/main.py diff --git a/tuner/learn_to_ask/main.py b/tuner/learn_to_ask/main.py index b286bba..dc71734 100644 --- a/tuner/learn_to_ask/main.py +++ b/tuner/learn_to_ask/main.py @@ -249,7 +249,6 @@ if __name__ == "__main__": temperature=1.0, tensor_parallel_size=1, inference_engine_num=4, - reasoning_parser=None, ) aux_models = { AUXILIARY_MODEL_NAME: TunerModelConfig( @@ -259,7 +258,6 @@ if __name__ == "__main__": temperature=0.7, tensor_parallel_size=2, inference_engine_num=1, - reasoning_parser=None, ), } algorithm = AlgorithmConfig(