Skip to content

fix: Enabled models list works#314

Merged
cdoern merged 1 commit intollamastack:mainfrom
gyliu513:models-list
Feb 5, 2026
Merged

fix: Enabled models list works#314
cdoern merged 1 commit intollamastack:mainfrom
gyliu513:models-list

Conversation

@gyliu513
Copy link
Copy Markdown
Contributor

Fixed #313

(llama-stack) (base) gualiu@gualiu-mac llamastack % llama-stack-client models list
INFO:httpx:HTTP Request: GET http://localhost:8321/v1/models "HTTP/1.1 200 OK"

Available Models

┏━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┓
┃ model_type       ┃ identifier                                                                  ┃ provider_resource_id                        ┃ metadata                                  ┃ provider_id                  ┃
┡━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┩
│ llm              │ ollama/all-minilm:latest                                                    │ all-minilm:latest                           │                                           │ ollama                       │
├──────────────────┼─────────────────────────────────────────────────────────────────────────────┼─────────────────────────────────────────────┼───────────────────────────────────────────┼──────────────────────────────┤
│ llm              │ ollama/llama3.2:1b-instruct-fp16                                            │ llama3.2:1b-instruct-fp16                   │                                           │ ollama                       │
├──────────────────┼─────────────────────────────────────────────────────────────────────────────┼─────────────────────────────────────────────┼───────────────────────────────────────────┼──────────────────────────────┤
│ llm              │ ollama/llama3.2-vision:latest                                               │ llama3.2-vision:latest                      │                                           │ ollama                       │
├──────────────────┼─────────────────────────────────────────────────────────────────────────────┼─────────────────────────────────────────────┼───────────────────────────────────────────┼──────────────────────────────┤
│ llm              │ ollama/llama3.2:3b-instruct-fp16                                            │ llama3.2:3b-instruct-fp16                   │                                           │ ollama                       │
├──────────────────┼─────────────────────────────────────────────────────────────────────────────┼─────────────────────────────────────────────┼───────────────────────────────────────────┼──────────────────────────────┤
│ llm              │ ollama/llama3.2:3b                                                          │ llama3.2:3b                                 │                                           │ ollama                       │
├──────────────────┼─────────────────────────────────────────────────────────────────────────────┼─────────────────────────────────────────────┼───────────────────────────────────────────┼──────────────────────────────┤
│ embedding        │ sentence-transformers/nomic-ai/nomic-embed-text-v1.5                        │ nomic-ai/nomic-embed-text-v1.5              │ {'embedding_dimension': 768}              │ sentence-transformers        │
└──────────────────┴─────────────────────────────────────────────────────────────────────────────┴─────────────────────────────────────────────┴───────────────────────────────────────────┴──────────────────────────────┘

Total models: 6

@meta-cla meta-cla bot added the cla signed label Jan 22, 2026
@cdoern cdoern merged commit acd5e64 into llamastack:main Feb 5, 2026
27 checks passed
@stainless-app stainless-app bot mentioned this pull request Feb 5, 2026
@gyliu513 gyliu513 deleted the models-list branch February 5, 2026 23:37
cdoern pushed a commit that referenced this pull request Mar 13, 2026
Automated Release PR
---


## 0.6.1-alpha.1 (2026-03-13)

Full Changelog:
[v0.5.0-alpha.2...v0.6.1-alpha.1](v0.5.0-alpha.2...v0.6.1-alpha.1)

### ⚠ BREAKING CHANGES

* improve consistency of post-training API endpoints

### Features

* accept list content blocks in Responses API function_call_output
([f6f1fc3](f6f1fc3))
* Add prompt_cache_key parameter support
([6b45699](6b45699))
* add skip_model_availability to openai_mixin for remote models
([7ef952b](7ef952b))
* add support for 'frequency_penalty' param to Responses API
([56d39cc](56d39cc))
* add support for 'presence_penalty' param to Responses API
([4f57d15](4f57d15))
* add support for /responses background parameter
([4f8bf45](4f8bf45))
* Add top_logprobs parameter support
([2196986](2196986))
* add top_p parameter support to responses API
([23e3b9f](23e3b9f))
* Add truncation parameter support
([7501365](7501365))
* improve consistency of post-training API endpoints
([99057fd](99057fd))
* **inference:** bidirectional reasoning token passthrough for chat
completions
([c314639](c314639))
* **vector_io:** Implement Contextual Retrieval for improved RAG search
quality
([89ec5a7](89ec5a7))


### Bug Fixes

* align chat completion usage schema with OpenAI spec
([3974d5d](3974d5d))
* Enabled models list works
([#314](#314))
([acd5e64](acd5e64))
* **inference:** use flat response message model for chat/completions
([e58e2e4](e58e2e4))
* **responses:** achieve full OpenResponses conformance — 6/6 tests
passing
([631ab2c](631ab2c))
* **stainless:** handle [DONE] SSE terminator in streaming responses
([17f0029](17f0029))
* **vector_io:** align Protocol signatures with request models
([ea58fd8](ea58fd8))


### Chores

* **api:** minor updates
([17a2705](17a2705))
* **ci:** bump uv version
([f014d4c](f014d4c))
* **ci:** skip uploading artifacts on stainless-internal branches
([dbddad9](dbddad9))
* **docs:** add missing descriptions
([f1a093b](f1a093b))
* format all `api.md` files
([0e3e262](0e3e262))
* **internal:** add request options to SSE classes
([2ecc682](2ecc682))
* **internal:** bump dependencies
([612291e](612291e))
* **internal:** fix lint error on Python 3.14
([a0f6975](a0f6975))
* **internal:** make `test_proxy_environment_variables` more resilient
([6bc2bb4](6bc2bb4))
* **internal:** make `test_proxy_environment_variables` more resilient
to env
([44bbae1](44bbae1))
* **test:** do not count install time for mock server timeout
([185de33](185de33))
* update mock server docs
([92cb087](92cb087))
* update placeholder string
([406b9bb](406b9bb))


### Refactors

* **types:** use `extra_items` from PEP 728
([629ca09](629ca09))

---
This pull request is managed by Stainless's [GitHub
App](https://github.com/apps/stainless-app).

The [semver version
number](https://semver.org/#semantic-versioning-specification-semver) is
based on included [commit
messages](https://www.conventionalcommits.org/en/v1.0.0/).
Alternatively, you can manually set the version number in the title of
this pull request.

For a better experience, it is recommended to use either rebase-merge or
squash-merge when merging this pull request.

🔗 Stainless [website](https://www.stainlessapi.com)
📚 Read the [docs](https://app.stainlessapi.com/docs)
🙋 [Reach out](mailto:support@stainlessapi.com) for help or questions

---------

Co-authored-by: stainless-app[bot] <142633134+stainless-app[bot]@users.noreply.github.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

llama-stack-client models list failed

2 participants