Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
1 change: 0 additions & 1 deletion .github/workflows/test-external-providers.yml
Original file line number Diff line number Diff line change
Expand Up @@ -72,7 +72,6 @@ jobs:
# temporary hack for file writing that should be done by the pip setup script
# https://github.com/containers/ramalama-stack/issues/53
mkdir -p ~/.llama/distributions/ramalama/
cp -r $GITHUB_WORKSPACE/src/ramalama_stack/providers.d/ ~/.llama/
cp $GITHUB_WORKSPACE/src/ramalama_stack/ramalama-run.yaml ~/.llama/distributions/ramalama/ramalama-run.yaml

- name: Run 'test-build.sh'
Expand Down
1 change: 0 additions & 1 deletion .github/workflows/test-lls-integration.yml
Original file line number Diff line number Diff line change
Expand Up @@ -56,7 +56,6 @@ jobs:
# temporary hack for file writing that should be done by the pip setup script
# https://github.com/containers/ramalama-stack/issues/53
mkdir -p ~/.llama/distributions/ramalama/
cp -r $GITHUB_WORKSPACE/src/ramalama_stack/providers.d/ ~/.llama/
cp $GITHUB_WORKSPACE/src/ramalama_stack/ramalama-run.yaml ~/.llama/distributions/ramalama/ramalama-run.yaml

- name: Run 'test-build.sh'
Expand Down
1 change: 0 additions & 1 deletion MANIFEST.in
Original file line number Diff line number Diff line change
@@ -1,2 +1 @@
recursive-include ramalama_stack/providers.d *
include ramalama_stack/ramalama-run.yaml
1 change: 0 additions & 1 deletion README.md
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,6 @@ This will install Llama Stack and RamaLama as well if they are not installed alr
> [!WARNING]
> The following workaround is currently needed to run this provider - see https://github.com/containers/ramalama-stack/issues/53 for more details
> ```bash
> curl --create-dirs --output ~/.llama/providers.d/remote/inference/ramalama.yaml https://raw.githubusercontent.com/containers/ramalama-stack/refs/tags/v0.2.5/src/ramalama_stack/providers.d/remote/inference/ramalama.yaml
> curl --create-dirs --output ~/.llama/distributions/ramalama/ramalama-run.yaml https://raw.githubusercontent.com/containers/ramalama-stack/refs/tags/v0.2.5/src/ramalama_stack/ramalama-run.yaml
> ```

Expand Down
8 changes: 5 additions & 3 deletions pyproject.toml
Original file line number Diff line number Diff line change
Expand Up @@ -21,8 +21,10 @@ dependencies = [
"datasets>=3.6.0",
"fastapi>=0.115.12",
"httpx>=0.28.1",
"llama-stack==0.2.14",
"llama-stack==0.6.0",
"llama-stack-client==0.6.0",
"mcp>=1.9.2",
"milvus-lite>=2.5.1",
"numpy>=2.2.6",
"openai>=1.82.0",
"opentelemetry-exporter-otlp-proto-http>=1.33.1",
Expand All @@ -31,7 +33,6 @@ dependencies = [
"psutil>=7.0.0",
"pydantic>=2.11.5",
"pymilvus>=2.5.10",
"ramalama==0.10.1",
"requests>=2.32.3",
"sentence-transformers>=3.0.0",
"six>=1.17.0",
Expand All @@ -40,6 +41,7 @@ dependencies = [
"trl>=0.18.1",
"urllib3>=2.4.0",
"uvicorn>=0.34.2",
"setuptools<70",
]

[dependency-groups]
Expand All @@ -62,7 +64,7 @@ package-dir = { "" = "src" }
include-package-data = true

[tool.setuptools.package-data]
"ramalama_stack" = ["providers.d/**/*", "ramalama-run.yaml"]
"ramalama_stack" = ["ramalama-run.yaml"]

[tool.ruff]
extend-exclude = ["*.ipynb"]
Loading