File size: 1,986 Bytes
28f6795
 
 
 
 
 
 
 
 
29317c8
ebf8c15
 
28f6795
 
 
 
 
 
2cda1f1
767f5c2
2cda1f1
28f6795
2cda1f1
 
28f6795
2cda1f1
28f6795
2cda1f1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
28f6795
2cda1f1
 
 
 
28f6795
 
2cda1f1
28f6795
 
2cda1f1
 
 
28f6795
 
2cda1f1
 
107e64a
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
---
license: apache-2.0
language:
- en
tags:
- control-vector
- creative-writing
base_model:
- zai-org/GLM-4.6
base_model_relation: adapter
pipeline_tag: text-generation
library_name: transformers
---

## gghfez/GLM-4.6-control-vectors

Creative Writing control-vectors for [zai-org/GLM-4.6](https://huggingface.co/zai-org/GLM-4.6)

Feedback is welcome and would be very helpful.

## Usage

Apply the *debias vector* and either the *positive* **or** *negative* vector when starting llama-server.
If both are applied, they will cancel each other out.

You can use either `--control-vector [/path/to/vector.gguf]` or --control-vector-scaled [/path/to/vector.gguf] [scale factor]

The *debias* vector must be set to `1.0`

IMPORTANT: The positive and negative axis control vectors must be used along with the relevant de-bias control vector - they cannot be used on their own!


### Llama.cpp / IK_Llama.cpp Example

**Creative writing**

```sh
llama-server --model GLM-4.6-UD-IQ2_XXS-00001-of-00003.gguf [your usual CLI arguments] \
    --control-vector-scaled glm-4.6_honesty_vs_machiavellianism__debias.gguf 1.0 \
    --control-vector-scaled glm-4.6_honesty_vs_machiavellianism__machiavellianism.gguf 1.0 \
```

**Creative Writing without reasoning**
```sh
llama-server --model GLM-4.6-UD-IQ2_XXS-00001-of-00003.gguf [your usual CLI arguments] \
    --chat-template-kwargs '{"enable_thinking": false}' \
    --control-vector-scaled glm-4.6_honesty_vs_machiavellianism__debias.gguf 1.0 \
    --control-vector-scaled glm-4.6_honesty_vs_machiavellianism__machiavellianism.gguf 1.0 \
```

**Assistant**

```sh
llama-server --model GLM-4.6-IQ3_KS-00001-of-00004.gguf [your usual CLI arguments] \
    --control-vector-scaled  glm-4.6_communication__debias.gguf 1.0 \
    --control-vector-scaled  glm-4.6_communication__direct_communication.gguf 1.0 \
```

### Limitations

With reasoning enabled on extreme quants like IQ2_XXS, very simple prompts like "Hi" may result in irrelevant replies.