mirror of
https://github.com/NousResearch/atropos.git
synced 2026-04-19 12:57:58 +00:00
Add n kwarg being ignored workaround
This commit is contained in:
parent
706097db21
commit
1aa72d7e7e
4 changed files with 208 additions and 4 deletions
23
atroposlib/tests/conftest.py
Normal file
23
atroposlib/tests/conftest.py
Normal file
|
|
@ -0,0 +1,23 @@
|
|||
import pytest
|
||||
|
||||
|
||||
def pytest_addoption(parser):
|
||||
parser.addoption(
|
||||
"--runproviders", action="store_true", default=False, help="run provider tests"
|
||||
)
|
||||
|
||||
|
||||
def pytest_configure(config):
|
||||
config.addinivalue_line(
|
||||
"markers", "providers: mark test as requires providers api keys to run"
|
||||
)
|
||||
|
||||
|
||||
def pytest_collection_modifyitems(config, items):
|
||||
if config.getoption("--runproviders"):
|
||||
# --runproviders given in cli: do not skip slow tests
|
||||
return
|
||||
skip_providers = pytest.mark.skip(reason="need --runproviders option to run")
|
||||
for item in items:
|
||||
if "providers" in item.keywords:
|
||||
item.add_marker(skip_providers)
|
||||
110
atroposlib/tests/test_openai_api_workarounds.py
Normal file
110
atroposlib/tests/test_openai_api_workarounds.py
Normal file
|
|
@ -0,0 +1,110 @@
|
|||
import asyncio
|
||||
import os
|
||||
|
||||
import dotenv
|
||||
import pytest
|
||||
|
||||
from atroposlib.envs.server_handling.openai_server import OpenaiConfig, OpenAIServer
|
||||
|
||||
|
||||
@pytest.mark.providers
|
||||
def test_openai_api_n_kwarg_ignore_discovery():
|
||||
dotenv.load_dotenv()
|
||||
openrouter_api_key = os.getenv("OPENROUTER_API_KEY")
|
||||
if not openrouter_api_key:
|
||||
pytest.skip("OPENROUTER_API_KEY not set")
|
||||
config = OpenaiConfig(
|
||||
api_key=openrouter_api_key,
|
||||
base_url="https://openrouter.ai/api/v1",
|
||||
model_name="openai/gpt-4.1-nano",
|
||||
timeout=1200,
|
||||
num_max_requests_at_once=512,
|
||||
num_requests_for_eval=64,
|
||||
rolling_buffer_length=1024,
|
||||
)
|
||||
assert not config.n_kwarg_is_ignored, "n kwarg is not ignored by default"
|
||||
n = 4
|
||||
server = OpenAIServer(
|
||||
config=config,
|
||||
)
|
||||
response = asyncio.run(
|
||||
server.chat_completion(
|
||||
messages=[
|
||||
{"role": "user", "content": "Hello, how are you?"},
|
||||
],
|
||||
n=n,
|
||||
)
|
||||
)
|
||||
assert server.config.n_kwarg_is_ignored, "n kwarg is should be set after discovery"
|
||||
print(len(response.choices), n)
|
||||
assert (
|
||||
len(response.choices) == n
|
||||
), f"Expected {n} responses, got {len(response.choices)}"
|
||||
|
||||
|
||||
@pytest.mark.providers
|
||||
def test_openai_api_n_kwarg_ignore_use():
|
||||
dotenv.load_dotenv()
|
||||
openrouter_api_key = os.getenv("OPENROUTER_API_KEY")
|
||||
if not openrouter_api_key:
|
||||
pytest.skip("OPENROUTER_API_KEY not set")
|
||||
config = OpenaiConfig(
|
||||
api_key=openrouter_api_key,
|
||||
base_url="https://openrouter.ai/api/v1",
|
||||
model_name="openai/gpt-4.1-nano",
|
||||
timeout=1200,
|
||||
num_max_requests_at_once=512,
|
||||
num_requests_for_eval=64,
|
||||
rolling_buffer_length=1024,
|
||||
n_kwarg_is_ignored=True,
|
||||
)
|
||||
server = OpenAIServer(
|
||||
config=config,
|
||||
)
|
||||
n = 4
|
||||
response = asyncio.run(
|
||||
server.chat_completion(
|
||||
messages=[
|
||||
{"role": "user", "content": "Hello, how are you?"},
|
||||
],
|
||||
n=n,
|
||||
)
|
||||
)
|
||||
assert server.config.n_kwarg_is_ignored, "n kwarg is should be set after discovery"
|
||||
assert (
|
||||
len(response.choices) == n
|
||||
), f"Expected {n} responses, got {len(response.choices)}"
|
||||
|
||||
|
||||
@pytest.mark.providers
|
||||
def test_openai_api_n_kwarg_supported():
|
||||
dotenv.load_dotenv()
|
||||
openai_api_key = os.getenv("OPENAI_API_KEY")
|
||||
if not openai_api_key:
|
||||
pytest.skip("OPENAI_API_KEY not set")
|
||||
config = OpenaiConfig(
|
||||
model_name="gpt-4.1-nano",
|
||||
timeout=1200,
|
||||
num_max_requests_at_once=512,
|
||||
num_requests_for_eval=64,
|
||||
rolling_buffer_length=1024,
|
||||
n_kwarg_is_ignored=False,
|
||||
)
|
||||
server = OpenAIServer(
|
||||
config=config,
|
||||
)
|
||||
n = 4
|
||||
response = asyncio.run(
|
||||
server.chat_completion(
|
||||
messages=[
|
||||
{"role": "user", "content": "Hello, how are you?"},
|
||||
],
|
||||
n=n,
|
||||
)
|
||||
)
|
||||
assert (
|
||||
not server.config.n_kwarg_is_ignored
|
||||
), "n kwarg should be used with supported models"
|
||||
assert (
|
||||
len(response.choices) == n
|
||||
), f"Expected {n} responses, got {len(response.choices)}"
|
||||
Loading…
Add table
Add a link
Reference in a new issue