diff --git a/example_trainer/grpo.py b/example_trainer/grpo.py index e6ba5e56..6fb0abed 100644 --- a/example_trainer/grpo.py +++ b/example_trainer/grpo.py @@ -432,8 +432,8 @@ def _load_model_with_lora(config: TrainingConfig) -> torch.nn.Module: # Default modules for most transformer models target_modules = ["q_proj", "v_proj"] - print(f"[Setup] Applying LoRA: r={config.lora_r}, alpha={config.lora_alpha}") - print(f"[Setup] Target modules: {target_modules}") + print(f"Applying LoRA: r={config.lora_r}, alpha={config.lora_alpha}") + print(f"Target modules: {target_modules}") lora_config = LoraConfig( task_type=TaskType.CAUSAL_LM, diff --git a/example_trainer/vllm_api_server.py b/example_trainer/vllm_api_server.py index 5db8e97d..9b43b838 100644 --- a/example_trainer/vllm_api_server.py +++ b/example_trainer/vllm_api_server.py @@ -1,6 +1,4 @@ -# Based on https://github.com/vllm-project/vllm/blob/main/vllm/entrypoints/api_server.py -# SPDX-License-Identifier: Apache-2.0 -# SPDX-FileCopyrightText: Copyright contributors to the vLLM project + """ Custom vLLM API server with weight bridge hooks for shared-memory training. @@ -232,7 +230,7 @@ async def _generate(request_dict: dict, raw_request: Request) -> Response: assert engine is not None results_generator = engine.generate(prompt, sampling_params, request_id) - # Streaming: yield results as they're generated + # Streaming: yield results as theyre generated async def stream_results() -> AsyncGenerator[bytes, None]: async for request_output in results_generator: prompt_text = request_output.prompt