@@ -21,22 +21,19 @@ class TestResponses:
21
21
22
22
@parametrize
23
23
def test_method_create_overload_1 (self , client : OpenAI ) -> None :
24
- response = client .responses .create (
25
- input = "string" ,
26
- model = "gpt-4o" ,
27
- )
24
+ response = client .responses .create ()
28
25
assert_matches_type (Response , response , path = ["response" ])
29
26
30
27
@parametrize
31
28
def test_method_create_with_all_params_overload_1 (self , client : OpenAI ) -> None :
32
29
response = client .responses .create (
33
- input = "string" ,
34
- model = "gpt-4o" ,
35
30
background = True ,
36
31
include = ["file_search_call.results" ],
32
+ input = "string" ,
37
33
instructions = "instructions" ,
38
34
max_output_tokens = 0 ,
39
35
metadata = {"foo" : "string" },
36
+ model = "gpt-4o" ,
40
37
parallel_tool_calls = True ,
41
38
previous_response_id = "previous_response_id" ,
42
39
prompt = {
@@ -72,10 +69,7 @@ def test_method_create_with_all_params_overload_1(self, client: OpenAI) -> None:
72
69
73
70
@parametrize
74
71
def test_raw_response_create_overload_1 (self , client : OpenAI ) -> None :
75
- http_response = client .responses .with_raw_response .create (
76
- input = "string" ,
77
- model = "gpt-4o" ,
78
- )
72
+ http_response = client .responses .with_raw_response .create ()
79
73
80
74
assert http_response .is_closed is True
81
75
assert http_response .http_request .headers .get ("X-Stainless-Lang" ) == "python"
@@ -84,10 +78,7 @@ def test_raw_response_create_overload_1(self, client: OpenAI) -> None:
84
78
85
79
@parametrize
86
80
def test_streaming_response_create_overload_1 (self , client : OpenAI ) -> None :
87
- with client .responses .with_streaming_response .create (
88
- input = "string" ,
89
- model = "gpt-4o" ,
90
- ) as http_response :
81
+ with client .responses .with_streaming_response .create () as http_response :
91
82
assert not http_response .is_closed
92
83
assert http_response .http_request .headers .get ("X-Stainless-Lang" ) == "python"
93
84
@@ -99,23 +90,21 @@ def test_streaming_response_create_overload_1(self, client: OpenAI) -> None:
99
90
@parametrize
100
91
def test_method_create_overload_2 (self , client : OpenAI ) -> None :
101
92
response_stream = client .responses .create (
102
- input = "string" ,
103
- model = "gpt-4o" ,
104
93
stream = True ,
105
94
)
106
95
response_stream .response .close ()
107
96
108
97
@parametrize
109
98
def test_method_create_with_all_params_overload_2 (self , client : OpenAI ) -> None :
110
99
response_stream = client .responses .create (
111
- input = "string" ,
112
- model = "gpt-4o" ,
113
100
stream = True ,
114
101
background = True ,
115
102
include = ["file_search_call.results" ],
103
+ input = "string" ,
116
104
instructions = "instructions" ,
117
105
max_output_tokens = 0 ,
118
106
metadata = {"foo" : "string" },
107
+ model = "gpt-4o" ,
119
108
parallel_tool_calls = True ,
120
109
previous_response_id = "previous_response_id" ,
121
110
prompt = {
@@ -151,8 +140,6 @@ def test_method_create_with_all_params_overload_2(self, client: OpenAI) -> None:
151
140
@parametrize
152
141
def test_raw_response_create_overload_2 (self , client : OpenAI ) -> None :
153
142
response = client .responses .with_raw_response .create (
154
- input = "string" ,
155
- model = "gpt-4o" ,
156
143
stream = True ,
157
144
)
158
145
@@ -163,8 +150,6 @@ def test_raw_response_create_overload_2(self, client: OpenAI) -> None:
163
150
@parametrize
164
151
def test_streaming_response_create_overload_2 (self , client : OpenAI ) -> None :
165
152
with client .responses .with_streaming_response .create (
166
- input = "string" ,
167
- model = "gpt-4o" ,
168
153
stream = True ,
169
154
) as response :
170
155
assert not response .is_closed
@@ -358,22 +343,19 @@ class TestAsyncResponses:
358
343
359
344
@parametrize
360
345
async def test_method_create_overload_1 (self , async_client : AsyncOpenAI ) -> None :
361
- response = await async_client .responses .create (
362
- input = "string" ,
363
- model = "gpt-4o" ,
364
- )
346
+ response = await async_client .responses .create ()
365
347
assert_matches_type (Response , response , path = ["response" ])
366
348
367
349
@parametrize
368
350
async def test_method_create_with_all_params_overload_1 (self , async_client : AsyncOpenAI ) -> None :
369
351
response = await async_client .responses .create (
370
- input = "string" ,
371
- model = "gpt-4o" ,
372
352
background = True ,
373
353
include = ["file_search_call.results" ],
354
+ input = "string" ,
374
355
instructions = "instructions" ,
375
356
max_output_tokens = 0 ,
376
357
metadata = {"foo" : "string" },
358
+ model = "gpt-4o" ,
377
359
parallel_tool_calls = True ,
378
360
previous_response_id = "previous_response_id" ,
379
361
prompt = {
@@ -409,10 +391,7 @@ async def test_method_create_with_all_params_overload_1(self, async_client: Asyn
409
391
410
392
@parametrize
411
393
async def test_raw_response_create_overload_1 (self , async_client : AsyncOpenAI ) -> None :
412
- http_response = await async_client .responses .with_raw_response .create (
413
- input = "string" ,
414
- model = "gpt-4o" ,
415
- )
394
+ http_response = await async_client .responses .with_raw_response .create ()
416
395
417
396
assert http_response .is_closed is True
418
397
assert http_response .http_request .headers .get ("X-Stainless-Lang" ) == "python"
@@ -421,10 +400,7 @@ async def test_raw_response_create_overload_1(self, async_client: AsyncOpenAI) -
421
400
422
401
@parametrize
423
402
async def test_streaming_response_create_overload_1 (self , async_client : AsyncOpenAI ) -> None :
424
- async with async_client .responses .with_streaming_response .create (
425
- input = "string" ,
426
- model = "gpt-4o" ,
427
- ) as http_response :
403
+ async with async_client .responses .with_streaming_response .create () as http_response :
428
404
assert not http_response .is_closed
429
405
assert http_response .http_request .headers .get ("X-Stainless-Lang" ) == "python"
430
406
@@ -436,23 +412,21 @@ async def test_streaming_response_create_overload_1(self, async_client: AsyncOpe
436
412
@parametrize
437
413
async def test_method_create_overload_2 (self , async_client : AsyncOpenAI ) -> None :
438
414
response_stream = await async_client .responses .create (
439
- input = "string" ,
440
- model = "gpt-4o" ,
441
415
stream = True ,
442
416
)
443
417
await response_stream .response .aclose ()
444
418
445
419
@parametrize
446
420
async def test_method_create_with_all_params_overload_2 (self , async_client : AsyncOpenAI ) -> None :
447
421
response_stream = await async_client .responses .create (
448
- input = "string" ,
449
- model = "gpt-4o" ,
450
422
stream = True ,
451
423
background = True ,
452
424
include = ["file_search_call.results" ],
425
+ input = "string" ,
453
426
instructions = "instructions" ,
454
427
max_output_tokens = 0 ,
455
428
metadata = {"foo" : "string" },
429
+ model = "gpt-4o" ,
456
430
parallel_tool_calls = True ,
457
431
previous_response_id = "previous_response_id" ,
458
432
prompt = {
@@ -488,8 +462,6 @@ async def test_method_create_with_all_params_overload_2(self, async_client: Asyn
488
462
@parametrize
489
463
async def test_raw_response_create_overload_2 (self , async_client : AsyncOpenAI ) -> None :
490
464
response = await async_client .responses .with_raw_response .create (
491
- input = "string" ,
492
- model = "gpt-4o" ,
493
465
stream = True ,
494
466
)
495
467
@@ -500,8 +472,6 @@ async def test_raw_response_create_overload_2(self, async_client: AsyncOpenAI) -
500
472
@parametrize
501
473
async def test_streaming_response_create_overload_2 (self , async_client : AsyncOpenAI ) -> None :
502
474
async with async_client .responses .with_streaming_response .create (
503
- input = "string" ,
504
- model = "gpt-4o" ,
505
475
stream = True ,
506
476
) as response :
507
477
assert not response .is_closed
0 commit comments