@@ -22,7 +22,7 @@ class TestBatches:
22
22
def test_method_create (self , client : OpenAI ) -> None :
23
23
batch = client .batches .create (
24
24
completion_window = "24h" ,
25
- endpoint = "/v1/chat/completions " ,
25
+ endpoint = "/v1/responses " ,
26
26
input_file_id = "string" ,
27
27
)
28
28
assert_matches_type (Batch , batch , path = ["response" ])
@@ -31,7 +31,7 @@ def test_method_create(self, client: OpenAI) -> None:
31
31
def test_method_create_with_all_params (self , client : OpenAI ) -> None :
32
32
batch = client .batches .create (
33
33
completion_window = "24h" ,
34
- endpoint = "/v1/chat/completions " ,
34
+ endpoint = "/v1/responses " ,
35
35
input_file_id = "string" ,
36
36
metadata = {"foo" : "string" },
37
37
)
@@ -41,7 +41,7 @@ def test_method_create_with_all_params(self, client: OpenAI) -> None:
41
41
def test_raw_response_create (self , client : OpenAI ) -> None :
42
42
response = client .batches .with_raw_response .create (
43
43
completion_window = "24h" ,
44
- endpoint = "/v1/chat/completions " ,
44
+ endpoint = "/v1/responses " ,
45
45
input_file_id = "string" ,
46
46
)
47
47
@@ -54,7 +54,7 @@ def test_raw_response_create(self, client: OpenAI) -> None:
54
54
def test_streaming_response_create (self , client : OpenAI ) -> None :
55
55
with client .batches .with_streaming_response .create (
56
56
completion_window = "24h" ,
57
- endpoint = "/v1/chat/completions " ,
57
+ endpoint = "/v1/responses " ,
58
58
input_file_id = "string" ,
59
59
) as response :
60
60
assert not response .is_closed
@@ -182,7 +182,7 @@ class TestAsyncBatches:
182
182
async def test_method_create (self , async_client : AsyncOpenAI ) -> None :
183
183
batch = await async_client .batches .create (
184
184
completion_window = "24h" ,
185
- endpoint = "/v1/chat/completions " ,
185
+ endpoint = "/v1/responses " ,
186
186
input_file_id = "string" ,
187
187
)
188
188
assert_matches_type (Batch , batch , path = ["response" ])
@@ -191,7 +191,7 @@ async def test_method_create(self, async_client: AsyncOpenAI) -> None:
191
191
async def test_method_create_with_all_params (self , async_client : AsyncOpenAI ) -> None :
192
192
batch = await async_client .batches .create (
193
193
completion_window = "24h" ,
194
- endpoint = "/v1/chat/completions " ,
194
+ endpoint = "/v1/responses " ,
195
195
input_file_id = "string" ,
196
196
metadata = {"foo" : "string" },
197
197
)
@@ -201,7 +201,7 @@ async def test_method_create_with_all_params(self, async_client: AsyncOpenAI) ->
201
201
async def test_raw_response_create (self , async_client : AsyncOpenAI ) -> None :
202
202
response = await async_client .batches .with_raw_response .create (
203
203
completion_window = "24h" ,
204
- endpoint = "/v1/chat/completions " ,
204
+ endpoint = "/v1/responses " ,
205
205
input_file_id = "string" ,
206
206
)
207
207
@@ -214,7 +214,7 @@ async def test_raw_response_create(self, async_client: AsyncOpenAI) -> None:
214
214
async def test_streaming_response_create (self , async_client : AsyncOpenAI ) -> None :
215
215
async with async_client .batches .with_streaming_response .create (
216
216
completion_window = "24h" ,
217
- endpoint = "/v1/chat/completions " ,
217
+ endpoint = "/v1/responses " ,
218
218
input_file_id = "string" ,
219
219
) as response :
220
220
assert not response .is_closed
0 commit comments