▶️Use Model
Last updated
Last updated
To help developers to easily find and run the right model, Nexa AI Hub provide comprehensive filter system and SDK.
The goal of Nexa Model Hub is to help developers find the most suitable models. To achieve the goal, we provide the following filter options:
Computer Vision
Image-to-Text
Image-to-Image
Audio
Text-to-Speech
Automatic Speech Recognition
Multimodal
Image-Text-to-text
NLP
Text Generation
Chat Completion
Question Answering
GGUF
GGUF is an optimized binary format designed for efficient model loading and saving, particularly suited for inference tasks. It is compatible with GGML and other executors. Developed by @ggerganov, the creator of llama.cpp (a widely-used C/C++ LLM inference framework), GGUF forms the foundation of the Nexa SDK's GGML component.
ONNX
ONNX is an open standard format for representing machine learning models. It establishes a common set of operators and a unified file format, enabling AI developers to utilize models across various frameworks, tools, runtimes, and compilers. ONNX shows unique performance advantages on devices with limited ram(mobile, IoT). The Nexa SDK's ONNX component is built upon the onnxruntime framework.
The Nexa Model Hub specializes in on-device models with parameter less than 10 billion.
This metric indicates the minimum random access memory (RAM) necessary for local model execution.
Displays the total storage space required for the model.
Follow the Installation to download the appropriate SDK for your operating system.
Nexa SDK enables developers to use one line of code to run the model that fits your specific requirement locally. The one line of code follows this pattern:
To see example and popular MODEL_PATH, see Supported Popular Models below:
Download more official models from Nexa model hub
To find the right code following "nexa", there are two ways:
Find the model in the model hub using search and filter, and click on "run this model" to copy the code to run model locally.
Follow the run a model section in CLI Reference
Model | Type | Format | Command |
---|---|---|---|
NLP
GGUF
nexa run octopus-v2
NLP
GGUF
nexa run octopus-v4
NLP
GGUF
nexa run tinyllama
NLP
GGUF/ONNX
nexa run llama2
NLP
GGUF/ONNX
nexa run llama3
NLP
GGUF/ONNX
nexa run llama3.1
NLP
GGUF/ONNX
nexa run gemma
NLP
GGUF
nexa run gemma2
NLP
GGUF
nexa run qwen1.5
NLP
GGUF/ONNX
nexa run qwen2
NLP
GGUF
nexa run qwen2.5
NLP
GGUF
nexa run mathqwen
NLP
GGUF/ONNX
nexa run mistral
NLP
GGUF
nexa run codegemma
NLP
GGUF
nexa run codellama
NLP
GGUF
nexa run codeqwen
NLP
GGUF
nexa run deepseek-coder
NLP
GGUF
nexa run dolphin-mistral
NLP
GGUF
nexa run phi2
NLP
GGUF/ONNX
nexa run phi3
NLP
GGUF
nexa run llama2-uncensored
NLP
GGUF
nexa run llama3-uncensored
NLP
GGUF
nexa run llama2-function-calling
Multimodal
GGUF
nexa run nanollava
Multimodal
GGUF
nexa run llava-phi3
Multimodal
GGUF
nexa run llava-llama3
Multimodal
GGUF
nexa run llava1.6-mistral
Multimodal
GGUF
nexa run llava1.6-vicuna
Computer Vision
GGUF
nexa run sd1-4
Computer Vision
GGUF/ONNX
nexa run sd1-5
Computer Vision
GGUF/ONNX
nexa run lcm-dreamshaper
Computer Vision
GGUF
nexa run hassaku-lcm
Computer Vision
GGUF
nexa run anything-lcm
Audio
BIN
nexa run faster-whisper-tiny
Audio
BIN
nexa run faster-whisper-small
Audio
BIN
nexa run faster-whisper-medium
Audio
BIN
nexa run faster-whisper-base
Audio
BIN
nexa run faster-whisper-large