Skip to content

HFerrahoglu/llm-neofetch-plus

Folders and files

NameName
Last commit message
Last commit date

Latest commit

ย 

History

5 Commits
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 

Repository files navigation

image

๐Ÿš€ LLM-Neofetch++

Version Python License Platform

Advanced System Information Tool for Local LLM Usage

Show detailed hardware specs optimized for running local AI models


โœจ Features

๐Ÿ” Comprehensive Hardware Detection

  • โœ… CPU: Model, cores, threads, frequency, temperature, usage
  • โœ… GPU: NVIDIA (nvidia-smi), AMD (rocm-smi), Intel Arc detection
  • โœ… VRAM: Total, used, and available video memory
  • โœ… RAM: Physical memory and swap information
  • โœ… Storage: Disk type (NVMe/SSD/HDD), capacity, speed benchmarks
  • โœ… Battery: Charge level, power status, time remaining (laptops)
  • โœ… Apple Silicon: M1/M2/M3/M4 detection with unified memory

๐ŸŽฏ Smart AI/LLM Features

  • ๐Ÿค– Model Recommendations: Personalized suggestions based on your hardware
  • ๐Ÿ“Š Quantization Guide: GGUF formats explained (Q2_K through Q8_0)
  • ๐Ÿš€ Backend Comparison: Ollama, llama.cpp, vLLM, ExLlamaV2, LM Studio
  • โšก Performance Estimates: Token/s predictions for different model sizes
  • ๐Ÿ’ก Optimization Tips: Specific advice for your system configuration

๐ŸŽจ Beautiful UI

  • ๐ŸŒˆ Color-coded Output: Easy to read with semantic colors
  • ๐Ÿ“Š Progress Bars: Visual representation of usage and capacity
  • ๐Ÿ”ง Configurable: Customize colors, emoji, detail level
  • ๐Ÿ“ฑ Responsive: Adapts to terminal width

๐Ÿ› ๏ธ Developer Friendly

  • ๐Ÿ“ค Export Formats: JSON, YAML, Markdown
  • ๐Ÿงช Unit Tests: Comprehensive test coverage
  • ๐Ÿ”Œ Modular Design: Easy to extend and customize
  • ๐Ÿ“ Type Hints: Full type annotations
  • ๐Ÿ› Verbose Mode: Detailed logging for debugging

๐Ÿ“ฆ Installation

From Source (Recommended)

# Clone the repository
git clone https://github.com/HFerrahoglu/llm-neofetch-plus.git
cd llm-neofetch-plus

# Install dependencies
pip install -r requirements.txt

# Run directly
python llm_neofetch.py

# Or install globally
pip install -e .
llm-neofetch

Using pip

pip install llm-neofetch-plus
llm-neofetch

๐ŸŽฎ Usage

Basic Usage

# Normal output (default)
llm-neofetch

# Minimal output
llm-neofetch -d 1

# Detailed output with all features
llm-neofetch -d 3

# Interactive mode (choose detail level)
llm-neofetch -i

Advanced Usage

# Run disk benchmark (takes ~10 seconds)
llm-neofetch -b

# Export to different formats
llm-neofetch --export report.json      # JSON format
llm-neofetch --export report.yaml      # YAML format
llm-neofetch --export report.md        # Markdown format

# Verbose logging for debugging
llm-neofetch -v

# Custom config file
llm-neofetch --config /path/to/config.yaml

# Combine options
llm-neofetch -d 3 -b --export full_report.json

๐Ÿ“ธ Screenshots

Normal Output

โ•”โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•—
โ•‘              โšก LLM โ€ข NEOFETCH ++  โšก                                   โ•‘
โ•‘         Advanced System Info for Local LLM Usage                         โ•‘
โ•‘                    v1.0.0 โ€ข 2026 Edition                                 โ•‘
โ•šโ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•โ•

โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
๐Ÿ’ป System Information
โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
  OS             Linux-6.5.0-1-amd64-x86_64-with-glibc2.38
  Kernel         6.5.0 (x86_64)
  Uptime         2d 14h 32m
  Python         3.11.5

โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
๐Ÿ”ง CPU
โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
  Model          AMD Ryzen 9 7950X 16-Core Processor
  Cores          16 physical / 32 threads
  Frequency      4200 MHz
  Usage          [โ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘]  35.2%

โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
๐ŸŽฎ GPU
โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
    ๐ŸŸข NVIDIA GeForce RTX 4090
      VRAM: 24.0 GB total
            [โ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘] 12.4/24.0 GB
      Usage          [โ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘โ–‘] 20.0%
      Temp: 58ยฐC

โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
๐ŸŽฏ Personalized Model Recommendations
โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€

  โ–ธ Extra Large Models (70-72B)
    โ€ข Llama 3.1 70B
    โ€ข Qwen2.5 72B

  โ–ธ Large Models (30-34B)
    โ€ข Llama 3.1 33B
    โ€ข Qwen2.5 32B

โš™๏ธ Configuration

LLM-Neofetch++ uses a YAML configuration file. By default, it looks for:

  1. ./config/config.yaml (in the project directory)
  2. ~/.config/llm-neofetch/config.yaml
  3. /etc/llm-neofetch/config.yaml

Sample Configuration

# UI Settings
ui:
  box_width: 76
  use_emoji: true
  show_progress_bars: true
  compact_mode: false

# Color Theme
colors:
  primary: "\033[1;34m"    # Blue
  success: "\033[1;32m"    # Green
  warning: "\033[1;33m"    # Yellow
  danger: "\033[1;31m"     # Red

# Performance Thresholds
thresholds:
  vram:
    excellent: 24  # GB
    good: 12
    moderate: 8

๐Ÿ”ง Development

Project Structure

llm-neofetch-plus/
โ”œโ”€โ”€ llm_neofetch.py          # Main application
โ”œโ”€โ”€ src/
โ”‚   โ”œโ”€โ”€ detectors.py         # Hardware detection modules
โ”‚   โ””โ”€โ”€ ui.py                # UI rendering and formatting
โ”œโ”€โ”€ config/
โ”‚   โ””โ”€โ”€ config.yaml          # Configuration file
โ”œโ”€โ”€ tests/
โ”‚   โ””โ”€โ”€ test_all.py          # Unit tests
โ”œโ”€โ”€ requirements.txt         # Python dependencies
โ”œโ”€โ”€ setup.py                 # Package setup
โ””โ”€โ”€ README.md               # This file

Running Tests

# Run all tests
python tests/test_all.py

# Run with pytest (if installed)
pytest tests/

# Run with coverage
pytest --cov=src tests/

Contributing

Contributions are welcome! Please:

  1. Fork the repository
  2. Create a feature branch (git checkout -b feature/amazing-feature)
  3. Commit your changes (git commit -m 'Add amazing feature')
  4. Push to the branch (git push origin feature/amazing-feature)
  5. Open a Pull Request

๐ŸŽฏ Use Cases

For AI/ML Developers

  • Quickly assess if your hardware can run specific models
  • Get token/s estimates before downloading large models
  • Understand which quantization format to use
  • Optimize your LLM stack configuration

For System Administrators

  • Monitor system resources for AI workloads
  • Export reports for documentation
  • Benchmark storage performance for model loading
  • Track GPU utilization and temperatures

For Researchers

  • Document hardware specs in papers
  • Compare performance across different systems
  • Generate reproducible system reports
  • Share hardware configurations

๐Ÿš€ Roadmap

  • Docker container support
  • Web dashboard (optional)
  • Historical tracking and graphs
  • Cloud GPU detection (AWS, GCP, Azure)
  • LLM benchmarking suite
  • Automatic model download suggestions
  • Integration with popular LLM frameworks

๐Ÿค Acknowledgments

  • Built with psutil for cross-platform system info
  • Inspired by neofetch
  • Community feedback from r/LocalLLaMA

๐Ÿ“„ License

MIT License - see LICENSE file for details


๐ŸŒŸ Star History

If you find this tool useful, please consider giving it a star โญ


๐Ÿ“ž Support


Made with โค๏ธ for the Local LLM Community

About

LLM-Neofetch++ is an advanced system information tool designed specifically for local LLM (Large Language Model) usage. It provides detailed hardware detection with personalized recommendations for running AI models on your system.

Topics

Resources

License

Stars

Watchers

Forks

Packages

 
 
 

Contributors

Languages