Quantization made by Richard Erkhov. [Github](https://github.com/RichardErkhov) [Discord](https://discord.gg/pvy7H8DZMG) [Request more models](https://github.com/RichardErkhov/quant_request) deepseek-coder-1.3b-typescript - GGUF - Model creator: https://huggingface.co/CodeGPTPlus/ - Original model: https://huggingface.co/CodeGPTPlus/deepseek-coder-1.3b-typescript/ | Name | Quant method | Size | | ---- | ---- | ---- | | [deepseek-coder-1.3b-typescript.Q2_K.gguf](https://huggingface.co/RichardErkhov/CodeGPTPlus_-_deepseek-coder-1.3b-typescript-gguf/blob/main/deepseek-coder-1.3b-typescript.Q2_K.gguf) | Q2_K | 0.52GB | | [deepseek-coder-1.3b-typescript.IQ3_XS.gguf](https://huggingface.co/RichardErkhov/CodeGPTPlus_-_deepseek-coder-1.3b-typescript-gguf/blob/main/deepseek-coder-1.3b-typescript.IQ3_XS.gguf) | IQ3_XS | 0.57GB | | [deepseek-coder-1.3b-typescript.IQ3_S.gguf](https://huggingface.co/RichardErkhov/CodeGPTPlus_-_deepseek-coder-1.3b-typescript-gguf/blob/main/deepseek-coder-1.3b-typescript.IQ3_S.gguf) | IQ3_S | 0.6GB | | [deepseek-coder-1.3b-typescript.Q3_K_S.gguf](https://huggingface.co/RichardErkhov/CodeGPTPlus_-_deepseek-coder-1.3b-typescript-gguf/blob/main/deepseek-coder-1.3b-typescript.Q3_K_S.gguf) | Q3_K_S | 0.6GB | | [deepseek-coder-1.3b-typescript.IQ3_M.gguf](https://huggingface.co/RichardErkhov/CodeGPTPlus_-_deepseek-coder-1.3b-typescript-gguf/blob/main/deepseek-coder-1.3b-typescript.IQ3_M.gguf) | IQ3_M | 0.63GB | | [deepseek-coder-1.3b-typescript.Q3_K.gguf](https://huggingface.co/RichardErkhov/CodeGPTPlus_-_deepseek-coder-1.3b-typescript-gguf/blob/main/deepseek-coder-1.3b-typescript.Q3_K.gguf) | Q3_K | 0.66GB | | [deepseek-coder-1.3b-typescript.Q3_K_M.gguf](https://huggingface.co/RichardErkhov/CodeGPTPlus_-_deepseek-coder-1.3b-typescript-gguf/blob/main/deepseek-coder-1.3b-typescript.Q3_K_M.gguf) | Q3_K_M | 0.66GB | | [deepseek-coder-1.3b-typescript.Q3_K_L.gguf](https://huggingface.co/RichardErkhov/CodeGPTPlus_-_deepseek-coder-1.3b-typescript-gguf/blob/main/deepseek-coder-1.3b-typescript.Q3_K_L.gguf) | Q3_K_L | 0.69GB | | [deepseek-coder-1.3b-typescript.IQ4_XS.gguf](https://huggingface.co/RichardErkhov/CodeGPTPlus_-_deepseek-coder-1.3b-typescript-gguf/blob/main/deepseek-coder-1.3b-typescript.IQ4_XS.gguf) | IQ4_XS | 0.7GB | | [deepseek-coder-1.3b-typescript.Q4_0.gguf](https://huggingface.co/RichardErkhov/CodeGPTPlus_-_deepseek-coder-1.3b-typescript-gguf/blob/main/deepseek-coder-1.3b-typescript.Q4_0.gguf) | Q4_0 | 0.72GB | | [deepseek-coder-1.3b-typescript.IQ4_NL.gguf](https://huggingface.co/RichardErkhov/CodeGPTPlus_-_deepseek-coder-1.3b-typescript-gguf/blob/main/deepseek-coder-1.3b-typescript.IQ4_NL.gguf) | IQ4_NL | 0.73GB | | [deepseek-coder-1.3b-typescript.Q4_K_S.gguf](https://huggingface.co/RichardErkhov/CodeGPTPlus_-_deepseek-coder-1.3b-typescript-gguf/blob/main/deepseek-coder-1.3b-typescript.Q4_K_S.gguf) | Q4_K_S | 0.76GB | | [deepseek-coder-1.3b-typescript.Q4_K.gguf](https://huggingface.co/RichardErkhov/CodeGPTPlus_-_deepseek-coder-1.3b-typescript-gguf/blob/main/deepseek-coder-1.3b-typescript.Q4_K.gguf) | Q4_K | 0.81GB | | [deepseek-coder-1.3b-typescript.Q4_K_M.gguf](https://huggingface.co/RichardErkhov/CodeGPTPlus_-_deepseek-coder-1.3b-typescript-gguf/blob/main/deepseek-coder-1.3b-typescript.Q4_K_M.gguf) | Q4_K_M | 0.81GB | | [deepseek-coder-1.3b-typescript.Q4_1.gguf](https://huggingface.co/RichardErkhov/CodeGPTPlus_-_deepseek-coder-1.3b-typescript-gguf/blob/main/deepseek-coder-1.3b-typescript.Q4_1.gguf) | Q4_1 | 0.8GB | | [deepseek-coder-1.3b-typescript.Q5_0.gguf](https://huggingface.co/RichardErkhov/CodeGPTPlus_-_deepseek-coder-1.3b-typescript-gguf/blob/main/deepseek-coder-1.3b-typescript.Q5_0.gguf) | Q5_0 | 0.87GB | | [deepseek-coder-1.3b-typescript.Q5_K_S.gguf](https://huggingface.co/RichardErkhov/CodeGPTPlus_-_deepseek-coder-1.3b-typescript-gguf/blob/main/deepseek-coder-1.3b-typescript.Q5_K_S.gguf) | Q5_K_S | 0.89GB | | [deepseek-coder-1.3b-typescript.Q5_K.gguf](https://huggingface.co/RichardErkhov/CodeGPTPlus_-_deepseek-coder-1.3b-typescript-gguf/blob/main/deepseek-coder-1.3b-typescript.Q5_K.gguf) | Q5_K | 0.93GB | | [deepseek-coder-1.3b-typescript.Q5_K_M.gguf](https://huggingface.co/RichardErkhov/CodeGPTPlus_-_deepseek-coder-1.3b-typescript-gguf/blob/main/deepseek-coder-1.3b-typescript.Q5_K_M.gguf) | Q5_K_M | 0.93GB | | [deepseek-coder-1.3b-typescript.Q5_1.gguf](https://huggingface.co/RichardErkhov/CodeGPTPlus_-_deepseek-coder-1.3b-typescript-gguf/blob/main/deepseek-coder-1.3b-typescript.Q5_1.gguf) | Q5_1 | 0.95GB | | [deepseek-coder-1.3b-typescript.Q6_K.gguf](https://huggingface.co/RichardErkhov/CodeGPTPlus_-_deepseek-coder-1.3b-typescript-gguf/blob/main/deepseek-coder-1.3b-typescript.Q6_K.gguf) | Q6_K | 1.09GB | | [deepseek-coder-1.3b-typescript.Q8_0.gguf](https://huggingface.co/RichardErkhov/CodeGPTPlus_-_deepseek-coder-1.3b-typescript-gguf/blob/main/deepseek-coder-1.3b-typescript.Q8_0.gguf) | Q8_0 | 1.33GB | Original model description: --- license: other base_model: deepseek-ai/deepseek-coder-1.3b-base tags: - axolotl - generated_from_trainer model-index: - name: deepseek-coder-1.3b-typescript results: [] datasets: - bigcode/the-stack-dedup widget: - text: "class Person {\n constructor(public name:" example_title: "class" - text: "function quickSort" example_title: "function" ---
[CodeGPT.co] | [š¦ Ollama] | [Discord] | [VSCode Extension]
### Fill In the Middle (FIM)
```python
<ļ½fimābeginļ½>function quickSort(arr: number[]): number[] {
if (arr.length <= 1) {
return arr;
}
const pivot = arr[0];
const left = [];
const right = [];
<ļ½fimāholeļ½>
return [...quickSort(left), pivot, ...quickSort(right)];
}<ļ½fimāendļ½>
```
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 20
- eval_batch_size: 20
- seed: 42
- gradient_accumulation_steps: 2
- total_train_batch_size: 40
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-06
- lr_scheduler_type: cosine
- lr_scheduler_warmup_steps: 261
- num_epochs: 1
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:-----:|:---------------:|
| 1.0745 | 0.0 | 1 | 0.8681 |
| 1.2267 | 0.05 | 1308 | 0.8130 |
| 1.1594 | 0.1 | 2616 | 0.8018 |
| 0.7674 | 0.15 | 3924 | 0.7942 |
| 0.6443 | 0.2 | 5232 | 0.7889 |
| 0.9155 | 0.25 | 6540 | 0.7847 |
| 0.7501 | 0.3 | 7848 | 0.7819 |
| 0.8835 | 0.35 | 9156 | 0.7792 |
| 0.7261 | 0.4 | 10464 | 0.7769 |
| 0.9746 | 0.45 | 11772 | 0.7748 |
| 0.6884 | 0.5 | 13080 | 0.7734 |
| 0.6104 | 0.55 | 14388 | 0.7722 |
| 0.8876 | 0.6 | 15696 | 0.7710 |
| 0.9567 | 0.65 | 17004 | 0.7703 |
| 0.6915 | 0.7 | 18312 | 0.7696 |
| 0.8874 | 0.75 | 19620 | 0.7691 |
| 0.6124 | 0.8 | 20928 | 0.7686 |
| 0.8147 | 0.85 | 22236 | 0.7684 |
| 0.8021 | 0.9 | 23544 | 0.7683 |
| 0.8665 | 0.95 | 24852 | 0.7681 |
### Framework versions
- Transformers 4.37.0.dev0
- Pytorch 2.0.1+cu118
- Datasets 2.16.1
- Tokenizers 0.15.0