Skip to content

feat(api): update via SDK Studio #212

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 1 commit into from
Aug 7, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
46 changes: 33 additions & 13 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -32,11 +32,21 @@ client = PromptFoundry(
api_key=os.environ.get("PROMPT_FOUNDRY_API_KEY"),
)

model_parameters = client.prompts.get_parameters(
completion_create_response = client.completion.create(
id="1212121",
variables={"hello": "world"},
append_messages=[
{
"role": "user",
"content": [
{
"type": "TEXT",
"text": "What is the weather in Seattle, WA?",
}
],
}
],
)
print(model_parameters.parameters)
print(completion_create_response.message)
```

While you can provide a `api_key` keyword argument,
Expand All @@ -60,11 +70,21 @@ client = AsyncPromptFoundry(


async def main() -> None:
model_parameters = await client.prompts.get_parameters(
completion_create_response = await client.completion.create(
id="1212121",
variables={"hello": "world"},
append_messages=[
{
"role": "user",
"content": [
{
"type": "TEXT",
"text": "What is the weather in Seattle, WA?",
}
],
}
],
)
print(model_parameters.parameters)
print(completion_create_response.message)


asyncio.run(main())
Expand Down Expand Up @@ -97,7 +117,7 @@ from prompt_foundry_python_sdk import PromptFoundry
client = PromptFoundry()

try:
client.prompts.get_parameters(
client.completion.create(
id="1212121",
)
except prompt_foundry_python_sdk.APIConnectionError as e:
Expand Down Expand Up @@ -142,7 +162,7 @@ client = PromptFoundry(
)

# Or, configure per-request:
client.with_options(max_retries=5).prompts.get_parameters(
client.with_options(max_retries=5).completion.create(
id="1212121",
)
```
Expand All @@ -167,7 +187,7 @@ client = PromptFoundry(
)

# Override per-request:
client.with_options(timeout=5.0).prompts.get_parameters(
client.with_options(timeout=5.0).completion.create(
id="1212121",
)
```
Expand Down Expand Up @@ -208,13 +228,13 @@ The "raw" Response object can be accessed by prefixing `.with_raw_response.` to
from prompt_foundry_python_sdk import PromptFoundry

client = PromptFoundry()
response = client.prompts.with_raw_response.get_parameters(
response = client.completion.with_raw_response.create(
id="1212121",
)
print(response.headers.get('X-My-Header'))

prompt = response.parse() # get the object that `prompts.get_parameters()` would have returned
print(prompt)
completion = response.parse() # get the object that `completion.create()` would have returned
print(completion.message)
```

These methods return an [`APIResponse`](https://github.com/prompt-foundry/python-sdk/tree/main/src/prompt_foundry_python_sdk/_response.py) object.
Expand All @@ -228,7 +248,7 @@ The above interface eagerly reads the full response body when you make the reque
To stream the response body, use `.with_streaming_response` instead, which requires a context manager and only reads the response body once you call `.read()`, `.text()`, `.json()`, `.iter_bytes()`, `.iter_text()`, `.iter_lines()` or `.parse()`. In the async client, these are async methods.

```python
with client.prompts.with_streaming_response.get_parameters(
with client.completion.with_streaming_response.create(
id="1212121",
) as response:
print(response.headers.get("X-My-Header"))
Expand Down
28 changes: 16 additions & 12 deletions tests/test_client.py
Original file line number Diff line number Diff line change
Expand Up @@ -721,11 +721,13 @@ def test_parse_retry_after_header(self, remaining_retries: int, retry_after: str
@mock.patch("prompt_foundry_python_sdk._base_client.BaseClient._calculate_retry_timeout", _low_retry_timeout)
@pytest.mark.respx(base_url=base_url)
def test_retrying_timeout_errors_doesnt_leak(self, respx_mock: MockRouter) -> None:
respx_mock.post("/sdk/v1/prompts/1212121").mock(side_effect=httpx.TimeoutException("Test timeout error"))
respx_mock.post("/sdk/v1/prompts/1212121/completion").mock(
side_effect=httpx.TimeoutException("Test timeout error")
)

with pytest.raises(APITimeoutError):
self.client.post(
"/sdk/v1/prompts/1212121",
"/sdk/v1/prompts/1212121/completion",
body=cast(object, dict()),
cast_to=httpx.Response,
options={"headers": {RAW_RESPONSE_HEADER: "stream"}},
Expand All @@ -736,11 +738,11 @@ def test_retrying_timeout_errors_doesnt_leak(self, respx_mock: MockRouter) -> No
@mock.patch("prompt_foundry_python_sdk._base_client.BaseClient._calculate_retry_timeout", _low_retry_timeout)
@pytest.mark.respx(base_url=base_url)
def test_retrying_status_errors_doesnt_leak(self, respx_mock: MockRouter) -> None:
respx_mock.post("/sdk/v1/prompts/1212121").mock(return_value=httpx.Response(500))
respx_mock.post("/sdk/v1/prompts/1212121/completion").mock(return_value=httpx.Response(500))

with pytest.raises(APIStatusError):
self.client.post(
"/sdk/v1/prompts/1212121",
"/sdk/v1/prompts/1212121/completion",
body=cast(object, dict()),
cast_to=httpx.Response,
options={"headers": {RAW_RESPONSE_HEADER: "stream"}},
Expand All @@ -763,9 +765,9 @@ def retry_handler(_request: httpx.Request) -> httpx.Response:
return httpx.Response(500)
return httpx.Response(200)

respx_mock.post("/sdk/v1/prompts/1212121").mock(side_effect=retry_handler)
respx_mock.post("/sdk/v1/prompts/1212121/completion").mock(side_effect=retry_handler)

response = client.prompts.with_raw_response.get_parameters(id="1212121")
response = client.completion.with_raw_response.create(id="1212121")

assert response.retries_taken == failures_before_success

Expand Down Expand Up @@ -1440,11 +1442,13 @@ async def test_parse_retry_after_header(self, remaining_retries: int, retry_afte
@mock.patch("prompt_foundry_python_sdk._base_client.BaseClient._calculate_retry_timeout", _low_retry_timeout)
@pytest.mark.respx(base_url=base_url)
async def test_retrying_timeout_errors_doesnt_leak(self, respx_mock: MockRouter) -> None:
respx_mock.post("/sdk/v1/prompts/1212121").mock(side_effect=httpx.TimeoutException("Test timeout error"))
respx_mock.post("/sdk/v1/prompts/1212121/completion").mock(
side_effect=httpx.TimeoutException("Test timeout error")
)

with pytest.raises(APITimeoutError):
await self.client.post(
"/sdk/v1/prompts/1212121",
"/sdk/v1/prompts/1212121/completion",
body=cast(object, dict()),
cast_to=httpx.Response,
options={"headers": {RAW_RESPONSE_HEADER: "stream"}},
Expand All @@ -1455,11 +1459,11 @@ async def test_retrying_timeout_errors_doesnt_leak(self, respx_mock: MockRouter)
@mock.patch("prompt_foundry_python_sdk._base_client.BaseClient._calculate_retry_timeout", _low_retry_timeout)
@pytest.mark.respx(base_url=base_url)
async def test_retrying_status_errors_doesnt_leak(self, respx_mock: MockRouter) -> None:
respx_mock.post("/sdk/v1/prompts/1212121").mock(return_value=httpx.Response(500))
respx_mock.post("/sdk/v1/prompts/1212121/completion").mock(return_value=httpx.Response(500))

with pytest.raises(APIStatusError):
await self.client.post(
"/sdk/v1/prompts/1212121",
"/sdk/v1/prompts/1212121/completion",
body=cast(object, dict()),
cast_to=httpx.Response,
options={"headers": {RAW_RESPONSE_HEADER: "stream"}},
Expand All @@ -1485,8 +1489,8 @@ def retry_handler(_request: httpx.Request) -> httpx.Response:
return httpx.Response(500)
return httpx.Response(200)

respx_mock.post("/sdk/v1/prompts/1212121").mock(side_effect=retry_handler)
respx_mock.post("/sdk/v1/prompts/1212121/completion").mock(side_effect=retry_handler)

response = await client.prompts.with_raw_response.get_parameters(id="1212121")
response = await client.completion.with_raw_response.create(id="1212121")

assert response.retries_taken == failures_before_success