Skip to content
Draft
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
376 changes: 238 additions & 138 deletions client-sdks/stainless/openapi.yml

Large diffs are not rendered by default.

363 changes: 143 additions & 220 deletions docs/docs/api-openai/conformance.mdx

Large diffs are not rendered by default.

364 changes: 235 additions & 129 deletions docs/static/deprecated-llama-stack-spec.yaml

Large diffs are not rendered by default.

364 changes: 235 additions & 129 deletions docs/static/experimental-llama-stack-spec.yaml

Large diffs are not rendered by default.

376 changes: 238 additions & 138 deletions docs/static/llama-stack-spec.yaml

Large diffs are not rendered by default.

470 changes: 49 additions & 421 deletions docs/static/openai-coverage.json

Large diffs are not rendered by default.

376 changes: 238 additions & 138 deletions docs/static/stainless-llama-stack-spec.yaml

Large diffs are not rendered by default.

4 changes: 2 additions & 2 deletions src/llama_stack/core/routers/vector_io.py
Original file line number Diff line number Diff line change
Expand Up @@ -408,8 +408,8 @@ async def openai_list_vector_stores(

# Determine pagination info
has_more = len(all_stores) > limit
first_id = limited_stores[0].id if limited_stores else None
last_id = limited_stores[-1].id if limited_stores else None
first_id = limited_stores[0].id if limited_stores else ""
last_id = limited_stores[-1].id if limited_stores else ""

return VectorStoreListResponse(
data=limited_stores,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -529,11 +529,11 @@ async def openai_create_vector_store(
"id": vector_store_id,
"object": "vector_store",
"created_at": created_at,
"name": params.name,
"name": params.name or "",
"usage_bytes": 0,
"file_counts": file_counts.model_dump(),
"status": status,
"expires_after": params.expires_after,
"expires_after": params.expires_after.model_dump() if params.expires_after else None,
"expires_at": None,
"last_active_at": created_at,
"file_ids": [],
Expand Down Expand Up @@ -619,8 +619,8 @@ async def openai_list_vector_stores(

# Determine pagination info
has_more = len(all_stores) > limit
first_id = data[0].id if data else None
last_id = data[-1].id if data else None
first_id = data[0].id if data else ""
last_id = data[-1].id if data else ""

return VectorStoreListResponse(
data=data,
Expand Down Expand Up @@ -655,7 +655,11 @@ async def openai_update_vector_store(
if request.name is not None:
store_info["name"] = request.name
if request.expires_after is not None:
store_info["expires_after"] = request.expires_after
store_info["expires_after"] = (
request.expires_after.model_dump()
if hasattr(request.expires_after, "model_dump")
else request.expires_after
)
if request.metadata is not None:
store_info["metadata"] = request.metadata

Expand Down Expand Up @@ -915,6 +919,7 @@ async def openai_attach_file_to_vector_store(
chunking_strategy=chunking_strategy,
created_at=created_at,
status="in_progress",
usage_bytes=0,
vector_store_id=vector_store_id,
)

Expand Down Expand Up @@ -1166,8 +1171,8 @@ async def openai_list_files_in_vector_store(

# Determine pagination info
has_more = len(file_objects) > limit
first_id = limited_files[0].id if file_objects else None
last_id = limited_files[-1].id if file_objects else None
first_id = limited_files[0].id if file_objects else ""
last_id = limited_files[-1].id if file_objects else ""

return VectorStoreListFilesResponse(
data=limited_files,
Expand Down Expand Up @@ -1217,6 +1222,7 @@ async def openai_retrieve_vector_store_file_contents(
)
return VectorStoreFileContentResponse(
data=content,
has_more=False,
)

async def openai_update_vector_store_file(
Expand Down Expand Up @@ -1546,8 +1552,8 @@ async def openai_list_files_in_vector_store_file_batch(

# Determine pagination info
has_more = len(file_objects) > limit
first_id = limited_files[0].id if limited_files else None
last_id = limited_files[-1].id if limited_files else None
first_id = limited_files[0].id if limited_files else ""
last_id = limited_files[-1].id if limited_files else ""

return VectorStoreFilesListInBatchResponse(
data=limited_files,
Expand Down
Loading
Loading