Skip to content

kjh2159/llama.cpp

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

llama.cpp

llama

License: MIT Release Server

Manifesto / ggml / ops

LLM inference in C/C++

Recent API changes

Hot topics


Quick start (llama.cpp)

The quick start guide of the basic llama.cpp follows the original repository. The main focus of IGNITE for on-device inference is based on llama-cli, guided by the llama-completion.

Quick start (IGNITE)

Model download

python downloader.py

Through this file, you can download models, which are pre-selected to evaluate themselves on IGNITE. If there is no preferred model, you can download and run your own gguf models also.

Build (on-device)

cd scripts && sh build-android.sh && cd ..

Run (on-device)

chmod +x scripts-termux/run.sh
su -c "sh scripts-termux/run.sh"

Build (Linux)

cd scripts && sh build.sh && cd ..

Run (Linux)

./build/bin/ignite \
    -m models/qwen-1.5-0.5b-chat-q4k.gguf \
    -cnv \
    --temp 0 \
    --top-k 1 \
    --threads 1 \
    --output-path outputs/hotpot_0_0.csv \
    --json-path dataset/hotpot_qa_30.json

This will be filled up. Please wait.

About

No description, website, or topics provided.

Resources

License

Contributing

Security policy

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 977