SLM 1.0

SLM 1.0 is a specialized language model trained by NeuroBrain, optimized for structured output generation, JSON schema compliance, and tool calling capabilities.

Model Details

Model Description

SLM 1.0 is a language model specifically trained to excel at:

  • Structured Output: Generating well-formatted, structured responses

  • JSON Schema: Producing outputs that strictly adhere to JSON schemas

  • Tool Calling: Effectively utilizing and calling external tools and functions

This model has been trained by NeuroBrain to provide reliable, structured responses suitable for production applications requiring precise output formatting.

Model Specifications

  • Architecture: SLM1ForCausalLM

  • Model Type: Causal Language Model

  • Context Length: 32,768 tokens

  • Hidden Size: 1,536

  • Number of Layers: 28

  • Attention Heads: 12

  • Vocabulary Size: 151,936

Training Information

  • Trained by: NeuroBrain

  • Training Method: Trained for structured output, JSON schema compliance, and tool calling

Usage

Basic Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "sihab/slm-1.0"

tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)

# Example: Generate structured output
prompt = "Generate a JSON object with user information"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_length=512)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)

Structured Output Generation

SLM 1.0 is particularly effective when you need structured outputs:

prompt = """
Generate a JSON object following this schema:
{
  "name": "string",
  "age": "number",
  "email": "string"
}
"""

inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_length=512, temperature=0.7)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)

Tool Calling

The model is optimized for tool calling scenarios:

prompt = """
Available tools:
- get_weather(location: str)
- send_email(to: str, subject: str, body: str)

User request: Check the weather in Paris and send me an email with the result.
"""

inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_length=1024)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)

Model Performance

SLM 1.0 demonstrates strong performance in:

  • JSON schema compliance

  • Structured data generation

  • Tool calling accuracy

  • Function parameter extraction

Limitations

  • The model may occasionally require post-processing to ensure strict JSON compliance

  • Tool calling accuracy depends on the clarity of tool descriptions provided

  • Maximum context length is 32,768 tokens

Citation

If you use SLM 1.0 in your research or applications, please cite:

@misc{slm1.0,
  title={SLM 1.0: A Language Model for Structured Output and Tool Calling},
  author={NeuroBrain},
  year={2025},
  howpublished={\url{https://huggingface.co/sihab/slm-1.0}}
}

License

This model is licensed under the Apache 2.0 license.

Contact

For questions, issues, or contributions, please contact NeuroBrain.


Model trained by NeuroBrain

Downloads last month
4
Safetensors
Model size
2B params
Tensor type
F16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support