Update README.md
Browse files
README.md
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
# Qwen2.5-Vex-Python
|
2 |
|
3 |
-
Qwen2.5-Vex-Python is a fine-tuned version of the Qwen2.5-Coder-7B-Instruct model, optimized for enhanced performance in Python code generation and understanding.
|
4 |
|
5 |
## Model Overview
|
6 |
|
@@ -12,10 +12,6 @@ Qwen2.5-Vex-Python is a fine-tuned version of the Qwen2.5-Coder-7B-Instruct mode
|
|
12 |
|
13 |
The Qwen2.5 series, developed by Alibaba Cloud's Qwen team, is a collection of large language models designed for various tasks, including code generation. The 7B variant strikes a balance between performance and resource requirements, making it suitable for a wide range of applications. citeturn0search0
|
14 |
|
15 |
-
## Training Methodology
|
16 |
-
|
17 |
-
This model was fine-tuned using the Unsloth framework in conjunction with Hugging Face's TRL (Transformers Reinforcement Learning) library. This combination enabled a training process that was twice as fast as traditional methods, resulting in a more efficient model without compromising performance.
|
18 |
-
|
19 |
## Usage
|
20 |
|
21 |
To utilize Qwen2.5-Vex-Python for Python code generation or understanding tasks, you can load the model using the Hugging Face Transformers library. Ensure that you have the necessary dependencies installed and that your environment supports 8-bit quantized models.
|
@@ -26,14 +22,6 @@ For detailed instructions on loading and using Qwen2.5 models, refer to the [Qwe
|
|
26 |
|
27 |
This model is licensed under the Apache-2.0 License. You are free to use, modify, and distribute this model, provided that you comply with the terms of the license.
|
28 |
|
29 |
-
## Acknowledgements
|
30 |
-
|
31 |
-
Special thanks to the developers of the Unsloth framework and Hugging Face's TRL library for their contributions to the machine learning community.
|
32 |
-
|
33 |
-
For more information on the base model and its capabilities, visit the [Qwen2.5-Coder GitHub repository](https://github.com/QwenLM/Qwen2.5-Coder). citeturn0search0
|
34 |
-
|
35 |
-
For a comprehensive overview of the Qwen2.5 series, refer to the [Qwen2.5 Blog Post](https://qwenlm.github.io/blog/qwen2.5/). citeturn0search2
|
36 |
-
|
37 |
---
|
38 |
base_model: unsloth/qwen2.5-coder-7b-instruct-bnb-4bit
|
39 |
tags:
|
|
|
1 |
# Qwen2.5-Vex-Python
|
2 |
|
3 |
+
Qwen2.5-Vex-Python is a fine-tuned version of the Qwen2.5-Coder-7B-Instruct model, optimized for enhanced performance in Python code generation and understanding.
|
4 |
|
5 |
## Model Overview
|
6 |
|
|
|
12 |
|
13 |
The Qwen2.5 series, developed by Alibaba Cloud's Qwen team, is a collection of large language models designed for various tasks, including code generation. The 7B variant strikes a balance between performance and resource requirements, making it suitable for a wide range of applications. citeturn0search0
|
14 |
|
|
|
|
|
|
|
|
|
15 |
## Usage
|
16 |
|
17 |
To utilize Qwen2.5-Vex-Python for Python code generation or understanding tasks, you can load the model using the Hugging Face Transformers library. Ensure that you have the necessary dependencies installed and that your environment supports 8-bit quantized models.
|
|
|
22 |
|
23 |
This model is licensed under the Apache-2.0 License. You are free to use, modify, and distribute this model, provided that you comply with the terms of the license.
|
24 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
25 |
---
|
26 |
base_model: unsloth/qwen2.5-coder-7b-instruct-bnb-4bit
|
27 |
tags:
|