Locai L1-Large
Collection
Locai L1-Large model release โข 3 items โข Updated
โข 2
This is the updated version of the Locai L1-Large model to improve multi-turn conversation ability and with an increase in self-improvement data in the training mix.
Locai L1-Large is an open-source instruction-tuned model based on Qwen3 235B Instruct (2507), post-trained using our Forget-Me-Not framework. This framework combines experience replay and self-improvement to enhance performance whilst mitigating catastrophic forgetting.
For more details on the model training, please refer to our technical report.
pip install transformers torch accelerate
from transformers import AutoModelForCausalLM, AutoTokenizer
model_name = "locailabs/locai-l1-large"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(
model_name,
device_map="auto",
torch_dtype="auto"
)
messages = [
{"role": "user", "content": "Explain quantum entanglement in simple terms"}
]
text = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer([text], return_tensors="pt").to(model.device)
outputs = model.generate(
**inputs,
max_new_tokens=2048,
temperature=0.7,
top_k=20,
top_p=0.8
)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)
from vllm import LLM, SamplingParams
llm = LLM(model="locailabs/locai-l1-large-2011")
sampling_params = SamplingParams(
temperature=0.7,
top_k=20,
top_p=0.8,
)
prompts = [
"Explain quantum entanglement in simple terms."
]
outputs = llm.generate(prompts, sampling_params)
for output in outputs:
print(output.outputs[0].text)
The model was trained on a curated dataset combining:
Locai L1-Large has been developed with consideration for:
@misc{locai2025l1large,
title={Locai L1-Large: Self-Improving Language Models with Forget-Me-Not},
author={Locai Labs},
year={2025}
}
Apache 2.0