remove comments

This commit is contained in:
Jai Suphavadeeprasit 2026-03-13 13:29:47 -04:00
parent 994e9c287d
commit 322e7e6623
5 changed files with 2 additions and 53 deletions

View file

@ -230,28 +230,8 @@ def resolve_openai_configs(
f"Merged Dict: {openai_config_dict}"
) from e
if isinstance(default_server_configs, APIServerConfig):
server_configs = [final_openai_config]
elif isinstance(default_server_configs, list):
if isinstance(default_server_configs, list):
server_configs = [final_openai_config]
else:
logger.warning(
f"Unexpected type for default_server_configs: {type(default_server_configs)}. "
f"Proceeding with single OpenAI server configuration based on merged settings."
)
server_configs = [final_openai_config]
if isinstance(server_configs, list):
logger.warning(
"resolve_openai_configs: returning list of %s config(s), URLs: %s",
len(server_configs),
[c.base_url for c in server_configs],
)
else:
logger.warning(
"resolve_openai_configs: returning single %s (base_url=%s) — "
"ServerManager will use template mode!",
type(server_configs).__name__,
getattr(server_configs, "base_url", "N/A"),
)
return server_configs

View file

@ -235,30 +235,15 @@ class GSM8kEnv(BaseEnv):
)
async with self.server.managed_server(tokenizer=self.tokenizer) as managed:
logger.warning(
"gsm8k collect_trajectories start group_size=%s max_tokens=%s question_chars=%s",
self.config.group_size,
self.config.max_token_length,
len(item["question"]),
)
chat_completions = await managed.chat_completion(
messages=[{"role": "system", "content": system_prompt}, user_message],
n=self.config.group_size,
max_tokens=self.config.max_token_length,
temperature=1.0,
)
logger.warning(
"gsm8k collect_trajectories completion_received choices=%s",
len(chat_completions.choices),
)
state = managed.get_state()
nodes = state["nodes"]
logger.warning(
"gsm8k collect_trajectories managed_state_nodes=%s",
len(nodes),
)
to_score = list()
to_backlog = list()

View file

@ -573,7 +573,7 @@ python -m example_trainer.vllm_api_server # NOT direct vllm commands
### WandB Logging
```bash
```bash
--use-wandb \
--wandb-project "my-grpo-training" \
--wandb-group "hermes-8b-gsm8k"

View file

@ -7,7 +7,6 @@ Handles communication with the Atropos API server for:
- Batch retrieval
"""
import os
import time as _time
import requests
@ -102,10 +101,6 @@ def get_batch(url: str = "http://localhost:8000"):
"""
response = requests.get(
f"{url}/batch",
headers={
"X-Atropos-Client": "trainer",
"X-Atropos-Pid": str(os.getpid()),
},
timeout=10,
)
data = response.json()

View file

@ -296,17 +296,6 @@ async def _generate(request_dict: dict, raw_request: Request) -> Response:
if engine is None:
raise HTTPException(status_code=503, detail="Engine not initialized")
request_preview = {
"has_prompt": "prompt" in request_dict,
"n": request_dict.get("n"),
"max_tokens": request_dict.get("max_tokens"),
"temperature": request_dict.get("temperature"),
"top_p": request_dict.get("top_p"),
"logprobs": request_dict.get("logprobs"),
"prompt_logprobs": request_dict.get("prompt_logprobs"),
}
logger.info("POST /generate received %s", request_preview)
prompt = request_dict.pop("prompt")
stream = request_dict.pop("stream", False)
request_dict["output_kind"] = RequestOutputKind.FINAL_ONLY