How to use from
vLLM
Install from pip and serve model
# Install vLLM from pip:
pip install vllm
# Start the vLLM server:
vllm serve "oxyapi/oxy-1-small-GGUF"
# Call the server using curl (OpenAI-compatible API):
curl -X POST "http://localhost:8000/v1/chat/completions" \
	-H "Content-Type: application/json" \
	--data '{
		"model": "oxyapi/oxy-1-small-GGUF",
		"messages": [
			{
				"role": "user",
				"content": "What is the capital of France?"
			}
		]
	}'
Use Docker
docker model run hf.co/oxyapi/oxy-1-small-GGUF:
Quick Links

Oxy 1 Small

Model Details

License

This model is licensed under the Apache 2.0 License.

Citation

If you find Oxy 1 Small useful in your research or applications, please cite it as:

@misc{oxy1small2024,
  title={Oxy 1 Small: A Fine-Tuned Qwen2.5-14B-Instruct Model for Role-Play},
  author={Oxygen (oxyapi)},
  year={2024},
  howpublished={\url{https://huggingface.co/oxyapi/oxy-1-small}},
}
Downloads last month
329
GGUF
Model size
15B params
Architecture
qwen2
Hardware compatibility
Log In to add your hardware

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for oxyapi/oxy-1-small-GGUF

Base model

Qwen/Qwen2.5-14B
Quantized
(9)
this model

Collection including oxyapi/oxy-1-small-GGUF