McaTech commited on
Commit
dbccb99
·
verified ·
1 Parent(s): 6eb62fa

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +123 -0
README.md ADDED
@@ -0,0 +1,123 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ````yaml
2
+ ---
3
+ license: apache-2.0
4
+ language:
5
+ - en
6
+ library_name: llama.cpp
7
+ tags:
8
+ - gguf
9
+ - quantized
10
+ - int8
11
+ - offline-ai
12
+ - local-llm
13
+ - chatnonet
14
+ model_type: causal
15
+ inference: true
16
+ pipeline_tag: text-generation
17
+ ---
18
+
19
+ # Model Card for ChatNONET
20
+
21
+ **ChatNONET** is a family of **offline**, quantized large language models fine-tuned for **question answering** with **direct, concise answers**. Designed for local execution using `llama.cpp`, ChatNONET is available in multiple sizes and optimized for Android or Python-based environments.
22
+
23
+ ## Model Details
24
+
25
+ ### Model Description
26
+
27
+ ChatNONET is intended for lightweight offline use, particularly on local devices like mobile phones or single-board computers. The models have been **fine-tuned for direct-answer QA** and quantized to **int8 (q8_0)** using `llama.cpp`.
28
+
29
+ | Model Name | Base Model | Size |
30
+ |----------------------------------|--------------------|--------|
31
+ | ChatNONET-135m-tuned-q8_0.gguf | Smollm | 135M |
32
+ | ChatNONET-300m-tuned-q8_0.gguf | Smollm | 300M |
33
+ | ChatNONET-1B-tuned-q8_0.gguf | LLaMA 3.2 | 1B |
34
+ | ChatNONET-3B-tuned-q8_0.gguf | LLaMA 3.2 | 3B |
35
+
36
+ - **Developed by:** McaTech (Michael Cobol Agan)
37
+ - **Model type:** Causal decoder-only transformer
38
+ - **Languages:** English
39
+ - **License:** Apache 2.0
40
+ - **Finetuned from:**
41
+ - Smollm (135M, 300M variants)
42
+ - LLaMA 3.2 (1B, 3B variants)
43
+
44
+ ## Uses
45
+
46
+ ### Direct Use
47
+
48
+ - Offline QA chatbot
49
+ - Local assistants (no internet required)
50
+ - Embedded Android or Python apps
51
+
52
+ ### Downstream Use
53
+
54
+ - Try the **Android app**: [Download ChatNONET APK](https://drive.google.com/file/d/1-5Ozx_VsOUBS5_b4yS40MCaNZge_5_1f/view?usp=sharing)
55
+
56
+ ### Out-of-Scope Use
57
+
58
+ - Long-form text generation
59
+ - Tasks requiring real-time web access
60
+ - Creative storytelling or coding tasks
61
+
62
+ ## Bias, Risks, and Limitations
63
+
64
+ ChatNONET may reproduce biases present in its base models or fine-tuning data. Outputs should not be relied upon for sensitive or critical decisions.
65
+
66
+ ### Recommendations
67
+
68
+ - Validate important responses
69
+ - Choose model size based on your device capability
70
+ - Avoid over-reliance for personal or legal advice
71
+
72
+ ## How to Get Started with the Model
73
+
74
+ ```bash
75
+ # Clone llama.cpp and build it
76
+ git clone https://github.com/ggerganov/llama.cpp
77
+ cd llama.cpp
78
+ make
79
+
80
+ # Run the model
81
+ ./llama-run -m ./ChatNONET-300m-tuned-q8_0.gguf -p "What is gravity?"
82
+ ````
83
+
84
+ ## Training Details
85
+
86
+ * **Finetuning Goal:** Direct-answer question answering
87
+ * **Precision:** FP16 mixed precision
88
+ * **Frameworks:** PyTorch, Transformers, Bitsandbytes
89
+ * **Quantization:** int8 GGUF (`q8_0`) via `llama.cpp`
90
+
91
+ ## Evaluation
92
+
93
+ * Evaluated internally on short QA prompts
94
+ * Capable of direct factual or logical answers
95
+ * Larger models perform better on reasoning tasks
96
+
97
+ ## Technical Specifications
98
+
99
+ * **Architecture:**
100
+
101
+ * Smollm (135M, 300M)
102
+ * LLaMA 3.2 (1B, 3B)
103
+ * **Format:** GGUF
104
+ * **Quantization:** q8\_0 (int8)
105
+ * **Deployment:** Mobile (Android) and desktop via `llama.cpp`
106
+
107
+ ## Citation
108
+
109
+ ```bibtex
110
+ @misc{chatnonet2025,
111
+ title={ChatNONET: Offline Quantized Q&A Models},
112
+ author={Michael Cobol Agan},
113
+ year={2025},
114
+ note={\url{https://huggingface.co/your-model-repo}},
115
+ }
116
+ ```
117
+
118
+ ## Contact
119
+
120
+ * **Author:** Michael Cobol Agan (McaTech)
121
+ * **Website / Download App:** https://mcatech.odoo.com/innovation
122
+
123
+ ```