mirror of
https://github.com/NousResearch/atropos.git
synced 2026-04-19 12:57:58 +00:00
remove comments
This commit is contained in:
parent
994e9c287d
commit
322e7e6623
5 changed files with 2 additions and 53 deletions
|
|
@ -230,28 +230,8 @@ def resolve_openai_configs(
|
|||
f"Merged Dict: {openai_config_dict}"
|
||||
) from e
|
||||
|
||||
if isinstance(default_server_configs, APIServerConfig):
|
||||
server_configs = [final_openai_config]
|
||||
elif isinstance(default_server_configs, list):
|
||||
if isinstance(default_server_configs, list):
|
||||
server_configs = [final_openai_config]
|
||||
else:
|
||||
logger.warning(
|
||||
f"Unexpected type for default_server_configs: {type(default_server_configs)}. "
|
||||
f"Proceeding with single OpenAI server configuration based on merged settings."
|
||||
)
|
||||
server_configs = [final_openai_config]
|
||||
|
||||
if isinstance(server_configs, list):
|
||||
logger.warning(
|
||||
"resolve_openai_configs: returning list of %s config(s), URLs: %s",
|
||||
len(server_configs),
|
||||
[c.base_url for c in server_configs],
|
||||
)
|
||||
else:
|
||||
logger.warning(
|
||||
"resolve_openai_configs: returning single %s (base_url=%s) — "
|
||||
"ServerManager will use template mode!",
|
||||
type(server_configs).__name__,
|
||||
getattr(server_configs, "base_url", "N/A"),
|
||||
)
|
||||
return server_configs
|
||||
|
|
|
|||
|
|
@ -235,30 +235,15 @@ class GSM8kEnv(BaseEnv):
|
|||
)
|
||||
|
||||
async with self.server.managed_server(tokenizer=self.tokenizer) as managed:
|
||||
logger.warning(
|
||||
"gsm8k collect_trajectories start group_size=%s max_tokens=%s question_chars=%s",
|
||||
self.config.group_size,
|
||||
self.config.max_token_length,
|
||||
len(item["question"]),
|
||||
)
|
||||
|
||||
chat_completions = await managed.chat_completion(
|
||||
messages=[{"role": "system", "content": system_prompt}, user_message],
|
||||
n=self.config.group_size,
|
||||
max_tokens=self.config.max_token_length,
|
||||
temperature=1.0,
|
||||
)
|
||||
logger.warning(
|
||||
"gsm8k collect_trajectories completion_received choices=%s",
|
||||
len(chat_completions.choices),
|
||||
)
|
||||
|
||||
state = managed.get_state()
|
||||
nodes = state["nodes"]
|
||||
logger.warning(
|
||||
"gsm8k collect_trajectories managed_state_nodes=%s",
|
||||
len(nodes),
|
||||
)
|
||||
|
||||
to_score = list()
|
||||
to_backlog = list()
|
||||
|
|
|
|||
|
|
@ -573,7 +573,7 @@ python -m example_trainer.vllm_api_server # NOT direct vllm commands
|
|||
|
||||
### WandB Logging
|
||||
|
||||
```bash
|
||||
```bash
|
||||
--use-wandb \
|
||||
--wandb-project "my-grpo-training" \
|
||||
--wandb-group "hermes-8b-gsm8k"
|
||||
|
|
|
|||
|
|
@ -7,7 +7,6 @@ Handles communication with the Atropos API server for:
|
|||
- Batch retrieval
|
||||
"""
|
||||
|
||||
import os
|
||||
import time as _time
|
||||
|
||||
import requests
|
||||
|
|
@ -102,10 +101,6 @@ def get_batch(url: str = "http://localhost:8000"):
|
|||
"""
|
||||
response = requests.get(
|
||||
f"{url}/batch",
|
||||
headers={
|
||||
"X-Atropos-Client": "trainer",
|
||||
"X-Atropos-Pid": str(os.getpid()),
|
||||
},
|
||||
timeout=10,
|
||||
)
|
||||
data = response.json()
|
||||
|
|
|
|||
|
|
@ -296,17 +296,6 @@ async def _generate(request_dict: dict, raw_request: Request) -> Response:
|
|||
if engine is None:
|
||||
raise HTTPException(status_code=503, detail="Engine not initialized")
|
||||
|
||||
request_preview = {
|
||||
"has_prompt": "prompt" in request_dict,
|
||||
"n": request_dict.get("n"),
|
||||
"max_tokens": request_dict.get("max_tokens"),
|
||||
"temperature": request_dict.get("temperature"),
|
||||
"top_p": request_dict.get("top_p"),
|
||||
"logprobs": request_dict.get("logprobs"),
|
||||
"prompt_logprobs": request_dict.get("prompt_logprobs"),
|
||||
}
|
||||
logger.info("POST /generate received %s", request_preview)
|
||||
|
||||
prompt = request_dict.pop("prompt")
|
||||
stream = request_dict.pop("stream", False)
|
||||
request_dict["output_kind"] = RequestOutputKind.FINAL_ONLY
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue