3030from litserve .specs .openai import OpenAISpec , ChatMessage
3131
3232
33- @pytest .mark .asyncio ()
33+ @pytest .mark .asyncio
3434async def test_openai_spec (openai_request_data ):
3535 spec = OpenAISpec ()
3636 server = ls .LitServer (TestAPI (), spec = spec )
@@ -45,7 +45,7 @@ async def test_openai_spec(openai_request_data):
4545
4646
4747# OpenAIWithUsage
48- @pytest .mark .asyncio ()
48+ @pytest .mark .asyncio
4949@pytest .mark .parametrize (
5050 ("api" , "batch_size" ),
5151 [
@@ -72,7 +72,7 @@ async def test_openai_token_usage(api, batch_size, openai_request_data, openai_r
7272 assert result ["usage" ] == openai_response_data ["usage" ]
7373
7474
75- @pytest .mark .asyncio ()
75+ @pytest .mark .asyncio
7676async def test_openai_spec_with_image (openai_request_data_with_image ):
7777 server = ls .LitServer (TestAPI (), spec = OpenAISpec ())
7878 with wrap_litserve_start (server ) as server :
@@ -85,7 +85,7 @@ async def test_openai_spec_with_image(openai_request_data_with_image):
8585 ), "LitAPI predict response should match with the generated output"
8686
8787
88- @pytest .mark .asyncio ()
88+ @pytest .mark .asyncio
8989async def test_override_encode (openai_request_data ):
9090 server = ls .LitServer (TestAPIWithCustomEncode (), spec = OpenAISpec ())
9191 with wrap_litserve_start (server ) as server :
@@ -98,7 +98,7 @@ async def test_override_encode(openai_request_data):
9898 ), "LitAPI predict response should match with the generated output"
9999
100100
101- @pytest .mark .asyncio ()
101+ @pytest .mark .asyncio
102102async def test_openai_spec_with_tools (openai_request_data_with_tools ):
103103 spec = OpenAISpec ()
104104 server = ls .LitServer (TestAPIWithToolCalls (), spec = spec )
@@ -118,7 +118,7 @@ async def test_openai_spec_with_tools(openai_request_data_with_tools):
118118 ], "LitAPI predict response should match with the generated output"
119119
120120
121- @pytest .mark .asyncio ()
121+ @pytest .mark .asyncio
122122async def test_openai_spec_with_response_format (openai_request_data_with_response_format ):
123123 spec = OpenAISpec ()
124124 server = ls .LitServer (TestAPIWithStructuredOutput (), spec = spec )
@@ -148,7 +148,7 @@ def encode_response(self, output):
148148 return ChatMessage (role = "assistant" , content = "This is a generated output" )
149149
150150
151- @pytest .mark .asyncio ()
151+ @pytest .mark .asyncio
152152async def test_openai_spec_validation (openai_request_data ):
153153 server = ls .LitServer (IncorrectAPI1 (), spec = OpenAISpec ())
154154 with pytest .raises (ValueError , match = "predict is not a generator" ), wrap_litserve_start (server ) as server :
@@ -172,7 +172,7 @@ def predict(self, prompt, context):
172172 return
173173
174174
175- @pytest .mark .asyncio ()
175+ @pytest .mark .asyncio
176176async def test_oai_prepopulated_context (openai_request_data ):
177177 openai_request_data ["max_tokens" ] = 3
178178 spec = OpenAISpec ()
@@ -194,7 +194,7 @@ def predict(self, prompt):
194194 raise HTTPException (501 , "test LitAPI.predict error" )
195195
196196
197- @pytest .mark .asyncio ()
197+ @pytest .mark .asyncio
198198async def test_fail_http (openai_request_data ):
199199 server = ls .LitServer (WrongLitAPI (), spec = ls .OpenAISpec ())
200200 with wrap_litserve_start (server ) as server :
0 commit comments