Skip to content
Open
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
10 changes: 0 additions & 10 deletions docs/docs/providers/openai_responses_limitations.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -98,16 +98,6 @@ The `reasoning` object in the output of Responses works for inference providers

---

### Service Tier

**Status:** Not Implemented

**Issue:** [#3550](https://github.com/llamastack/llama-stack/issues/3550)

Responses has a field `service_tier` that can be used to prioritize access to inference resources. Not all inference providers have such a concept, but Llama Stack pass through this value for those providers that do. Currently it does not.

---

### Incomplete Details

**Status:** Not Implemented
Expand Down