modelId
string
author
string
last_modified
timestamp[us, tz=UTC]
downloads
int64
likes
int64
library_name
string
tags
list
pipeline_tag
string
createdAt
timestamp[us, tz=UTC]
card
string
SoloBSD/solobsd-7b
SoloBSD
2024-01-30T19:46:39Z
2
0
transformers
[ "transformers", "safetensors", "gguf", "text-generation-inference", "unsloth", "llama", "en", "base_model:unsloth/llama-2-7b-bnb-4bit", "base_model:quantized:unsloth/llama-2-7b-bnb-4bit", "license:apache-2.0", "endpoints_compatible", "region:us" ]
null
2024-01-30T17:56:49Z
--- language: - en license: apache-2.0 tags: - text-generation-inference - transformers - unsloth - llama - gguf base_model: unsloth/llama-2-7b-bnb-4bit --- # Uploaded model - **Developed by:** solobsd - **License:** apache-2.0 - **Finetuned from model :** unsloth/llama-2-7b-bnb-4bit This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
LoneStriker/CodeFuse-DeepSeek-33B-GGUF
LoneStriker
2024-01-30T19:30:14Z
84
22
null
[ "gguf", "license:other", "endpoints_compatible", "region:us" ]
null
2024-01-30T18:06:05Z
--- license: other tasks: - code-generation --- # Model Card for CodeFuse-DeepSeek-33B ![logo](LOGO.jpg) [[中文]](#chinese) [[English]](#english) <a id="english"></a> ## Model Description CodeFuse-DeepSeek-33B is a 33B Code-LLM finetuned by QLoRA on multiple code-related tasks on the base model DeepSeek-Coder-33B. <br> ## News and Updates 🔥🔥🔥 2024-01-12 CodeFuse-DeepSeek-33B has been released, achieving a pass@1 (greedy decoding) score of 78.65% on HumanEval. 🔥🔥🔥 2024-01-12 CodeFuse-Mixtral-8x7B has been released, achieving a pass@1 (greedy decoding) score of 56.1% on HumanEval, which is a 15% increase compared to Mixtral-8x7b's 40%. 🔥🔥 2023-11-10 CodeFuse-CodeGeeX2-6B has been released, achieving a pass@1 (greedy decoding) score of 45.12% on HumanEval, which is a 9.22% increase compared to CodeGeeX2 35.9%. 🔥🔥 2023-10-20 CodeFuse-QWen-14B technical documentation has been released. For those interested, please refer to the CodeFuse article on our WeChat official account via the provided link.(https://mp.weixin.qq.com/s/PCQPkvbvfxSPzsqjOILCDw) 🔥🔥 2023-10-16 CodeFuse-QWen-14B has been released, achieving a pass@1 (greedy decoding) score of 48.78% on HumanEval, which is a 16% increase compared to Qwen-14b's 32.3%. 🔥🔥 2023-09-27 CodeFuse-StarCoder-15B has been released, achieving a pass@1 (greedy decoding) score of 54.9% on HumanEval, which is a 21% increase compared to StarCoder's 33.6%. 🔥🔥 2023-09-26 We are pleased to announce the release of the 4-bit quantized version of CodeFuse-CodeLlama-34B. Despite the quantization process, the model still achieves a remarkable 73.8% accuracy (greedy decoding) on the HumanEval pass@1 metric. 🔥🔥 2023-09-11 CodeFuse-CodeLlama-34B has achieved 74.4% of pass@1 (greedy decoding) on HumanEval, which is SOTA results for openspurced LLMs at present. <br> ## Code Community **Homepage**: 🏡 https://github.com/codefuse-ai (**Please give us your support with a Star🌟 + Fork🚀 + Watch👀**) + If you wish to fine-tune the model yourself, you can visit ✨[MFTCoder](https://github.com/codefuse-ai/MFTCoder)✨✨ + If you wish to see a demo of the model, you can visit ✨[CodeFuse Demo](https://github.com/codefuse-ai/codefuse)✨✨ <br> ## Performance ### Code | Model | HumanEval(pass@1) | Date | |:----------------------------|:-----------------:|:-------:| | **CodeFuse-DeepSeek-33B** | **78.65%** | 2024.01 | | **CodeFuse-Mixtral-8x7B** | **56.10%** | 2024.01 | | **CodeFuse-CodeLlama-34B** | 74.4% | 2023.9 | |**CodeFuse-CodeLlama-34B-4bits** | 73.8% | 2023.9 | | **CodeFuse-StarCoder-15B** | 54.9% | 2023.9 | | **CodeFuse-QWen-14B** | 48.78% | 2023.10 | | **CodeFuse-CodeGeeX2-6B** | 45.12% | 2023.11 | | WizardCoder-Python-34B-V1.0 | 73.2% | 2023.8 | | GPT-4(zero-shot) | 67.0% | 2023.3 | | PanGu-Coder2 15B | 61.6% | 2023.8 | | CodeLlama-34b-Python | 53.7% | 2023.8 | | CodeLlama-34b | 48.8% | 2023.8 | | GPT-3.5(zero-shot) | 48.1% | 2022.11 | | OctoCoder | 46.2% | 2023.8 | | StarCoder-15B | 33.6% | 2023.5 | | Qwen-14b | 32.3% | 2023.10 | ### NLP ![NLP Performance Radar](codefuse-deepseek-33b-nlp.png) <br> ## Requirements * python>=3.8 * pytorch>=2.0.0 * transformers>=4.33.2 * Sentencepiece * CUDA 11.4 <br> ## Inference String Format The inference string is a concatenated string formed by combining conversation data(system, human and bot contents) in the training data format. It is used as input during the inference process. Here are examples of prompts used to request the model: **Multi-Round with System Prompt:** ```python """ <s>system System instruction <s>human Human 1st round input <s>bot Bot 1st round output<|end▁of▁sentence|> <s>human Human 2nd round input <s>bot Bot 2nd round output<|end▁of▁sentence|> ... ... ... <s>human Human nth round input <s>bot """ ``` **Single-Round without System Prompt:** ```python """ <s>human User prompt... <s>bot """ ``` In this format, the system section is optional and the conversation can be either single-turn or multi-turn. When applying inference, you always make your input string end with "\<s\>bot" to ask the model generating answers. For example, the format used to infer HumanEval is like the following: ``` <s>human # language: Python from typing import List def separate_paren_groups(paren_string: str) -> List[str]: """ Input to this function is a string containing multiple groups of nested parentheses. Your goal is to separate those group into separate strings and return the list of those. Separate groups are balanced (each open brace is properly closed) and not nested within each other Ignore any spaces in the input string. >>> separate_paren_groups('( ) (( )) (( )( ))') ['()', '(())', '(()())'] """ <s>bot ``` Specifically, we also add the Programming Language Tag (e.g. "```# language: Python```" for Python) used by CodeGeex models. ## Quickstart ```python import torch from transformers import AutoTokenizer, AutoModelForCausalLM, GenerationConfig model_dir = "codefuse-ai/CodeFuse-DeepSeek-33B" def load_model_tokenizer(model_path): tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True) tokenizer.eos_token = "<|end▁of▁sentence|>" tokenizer.pad_token = "<|end▁of▁sentence|>" tokenizer.eos_token_id = tokenizer.convert_tokens_to_ids(tokenizer.eos_token) tokenizer.pad_token_id = tokenizer.convert_tokens_to_ids(tokenizer.pad_token) tokenizer.padding_side = "left" model = AutoModelForCausalLM.from_pretrained(model_path, device_map='auto',torch_dtype=torch.bfloat16, trust_remote_code=True) return model, tokenizer HUMAN_ROLE_START_TAG = "<s>human\n" BOT_ROLE_START_TAG = "<s>bot\n" text_list = [f'{HUMAN_ROLE_START_TAG}Write a QuickSort program\n#Python\n{BOT_ROLE_START_TAG}'] model, tokenizer = load_model_tokenizer(model_dir) inputs = tokenizer(text_list, return_tensors='pt', padding=True, add_special_tokens=False).to('cuda') input_ids = inputs["input_ids"] attention_mask = inputs["attention_mask"] generation_config = GenerationConfig( eos_token_id=tokenizer.eos_token_id, pad_token_id=tokenizer.pad_token_id, temperature=0.1, max_new_tokens=512, num_return_sequences=1, num_beams=1, top_p=0.95, do_sample=False ) outputs = model.generate( inputs= input_ids, attention_mask=attention_mask, **generation_config.to_dict() ) gen_text = tokenizer.batch_decode(outputs[:, input_ids.shape[1]:], skip_special_tokens=True) print(gen_text[0]) ``` <a id="chinese"></a> ## 模型简介 CodeFuse-DeepSeek-33B 是一个通过QLoRA对基座模型DeepSeek-Coder-33B进行多代码任务微调而得到的代码大模型。 <br> ## 新闻 🔥🔥🔥 2024-01-12 CodeFuse-DeepSeek-33B模型发布,模型在HumanEval pass@1指标为78.65% (贪婪解码)。 🔥🔥🔥 2023-11-10 开源了CodeFuse-CodeGeeX2-6B模型,在HumanEval pass@1(greedy decoding)上可以达到48.12%, 比CodeGeeX2提高了9.22%的代码能力(HumanEval) 🔥🔥🔥 2023-10-20 公布了CodeFuse-QWen-14B技术文档,感兴趣详见微信公众号CodeFuse文章:https://mp.weixin.qq.com/s/PCQPkvbvfxSPzsqjOILCDw 🔥🔥🔥 2023-10-16开源了CodeFuse-QWen-14B模型,在HumanEval pass@1(greedy decoding)上可以达到48.78%, 比Qwen-14b提高了16%的代码能力(HumanEval) 🔥🔥🔥 2023-09-27开源了CodeFuse-StarCoder-15B模型,在HumanEval pass@1(greedy decoding)上可以达到54.9%, 比StarCoder提高了21%的代码能力(HumanEval) 🔥🔥🔥 2023-09-26 [CodeFuse-CodeLlama-34B 4bits](https://modelscope.cn/models/codefuse-ai/CodeFuse-CodeLlama-34B-4bits/summary)量化版本发布,量化后模型在HumanEval pass@1指标为73.8% (贪婪解码)。 🔥🔥🔥 2023-09-11 [CodeFuse-CodeLlama-34B](https://modelscope.cn/models/codefuse-ai/CodeFuse-CodeLlama-34B/summary)发布,HumanEval pass@1指标达到74.4% (贪婪解码), 为当前开源SOTA。 <br> ## 代码社区 **大本营**: 🏡 https://github.com/codefuse-ai (**请支持我们的项目Star🌟 + Fork🚀 + Watch👀**) + 如果您想自己微调该模型,可以访问 ✨[MFTCoder](https://github.com/codefuse-ai/MFTCoder)✨✨ + 如果您想观看该模型示例,可以访问 ✨[CodeFuse Demo](https://github.com/codefuse-ai/codefuse)✨✨ <br> ## 评测表现 ### 代码 | 模型 | HumanEval(pass@1) | 日期 | |:----------------------------|:-----------------:|:-------:| | **CodeFuse-CodeLlama-34B** | 74.4% | 2023.9 | |**CodeFuse-CodeLlama-34B-4bits** | 73.8% | 2023.9 | | WizardCoder-Python-34B-V1.0 | 73.2% | 2023.8 | | GPT-4(zero-shot) | 67.0% | 2023.3 | | PanGu-Coder2 15B | 61.6% | 2023.8 | | CodeLlama-34b-Python | 53.7% | 2023.8 | | CodeLlama-34b | 48.8% | 2023.8 | | GPT-3.5(zero-shot) | 48.1% | 2022.11 | | OctoCoder | 46.2% | 2023.8 | | StarCoder-15B | 33.6% | 2023.5 | | Qwen-14b | 32.3% | 2023.10 | | **CodeFuse-StarCoder-15B** | 54.9% | 2023.9 | | **CodeFuse-QWen-14B** | 48.78% | 2023.8 | | **CodeFuse-CodeGeeX2-6B** | 45.12% | 2023.11 | | **CodeFuse-DeepSeek-33B**. | **78.65%** | 2024.01 | ### NLP ![NLP Performance Radar](codefuse-deepseek-33b-nlp.png) ## Requirements * python>=3.8 * pytorch>=2.0.0 * transformers>=4.33.2 * Sentencepiece * CUDA 11.4 <br> ## 推理数据格式 推理数据为模型在训练数据格式下拼接的字符串形式,它也是推理时输入prompt拼接的方式. 下面分别是带系统提示的多轮会话格式和不带系统提示的单轮会话格式: **带System提示的多轮会话格式:** ```python """ <s>system System instruction <s>human Human 1st round input <s>bot Bot 1st round output<|end▁of▁sentence|> <s>human Human 2nd round input <s>bot Bot 2nd round output<|end▁of▁sentence|> ... ... ... <s>human Human nth round input <s>bot """ ``` **不带System提示的单轮会话格式:** ```python """ <s>human User prompt... <s>bot """ ``` 在这个格式中,System提示是可选的(按需设定),支持单轮会话也支持多轮会话。推理时,请确保拼接的prompt字符串以"\<s\>bot\n"结尾,引导模型生成回答。 例如,推理HumanEval数据时使用的格式如下所示: ```python <s>human # language: Python from typing import List def separate_paren_groups(paren_string: str) -> List[str]: """ Input to this function is a string containing multiple groups of nested parentheses. Your goal is to separate those group into separate strings and return the list of those. Separate groups are balanced (each open brace is properly closed) and not nested within each other Ignore any spaces in the input string. >>> separate_paren_groups('( ) (( )) (( )( ))') ['()', '(())', '(()())'] """ <s>bot ``` 特别地,我们也使用了CodeGeeX系列模型采用的编程语言区分标签(例如,对于Python语言,我们会使用"```# language: Python```")。 ## 快速使用 ```python import torch from transformers import AutoTokenizer, AutoModelForCausalLM, GenerationConfig model_dir = "codefuse-ai/CodeFuse-DeepSeek-33B" def load_model_tokenizer(model_path): tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True) tokenizer.eos_token = "<|end▁of▁sentence|>" tokenizer.pad_token = "<|end▁of▁sentence|>" tokenizer.eos_token_id = tokenizer.convert_tokens_to_ids(tokenizer.eos_token) tokenizer.pad_token_id = tokenizer.convert_tokens_to_ids(tokenizer.pad_token) tokenizer.padding_side = "left" model = AutoModelForCausalLM.from_pretrained(model_path, device_map='auto',torch_dtype=torch.bfloat16, trust_remote_code=True) return model, tokenizer HUMAN_ROLE_START_TAG = "<s>human\n" BOT_ROLE_START_TAG = "<s>bot\n" text_list = [f'{HUMAN_ROLE_START_TAG}请写一个快排程序\n#Python\n{BOT_ROLE_START_TAG}'] model, tokenizer = load_model_tokenizer(model_dir) inputs = tokenizer(text_list, return_tensors='pt', padding=True, add_special_tokens=False).to('cuda') input_ids = inputs["input_ids"] attention_mask = inputs["attention_mask"] generation_config = GenerationConfig( eos_token_id=tokenizer.eos_token_id, pad_token_id=tokenizer.pad_token_id, temperature=0.2, max_new_tokens=512, num_return_sequences=1, num_beams=1, top_p=0.95, do_sample=False ) outputs = model.generate( inputs= input_ids, attention_mask=attention_mask, **generation_config.to_dict() ) gen_text = tokenizer.batch_decode(outputs[:, input_ids.shape[1]:], skip_special_tokens=True) print(gen_text[0]) ```
birgermoell/NeuralBeagle-Flashback
birgermoell
2024-01-30T19:21:46Z
8
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "merge", "mergekit", "lazymergekit", "mlabonne/NeuralBeagle14-7B", "base_model:mlabonne/NeuralBeagle14-7B", "base_model:finetune:mlabonne/NeuralBeagle14-7B", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-01-30T19:09:51Z
--- tags: - merge - mergekit - lazymergekit - mlabonne/NeuralBeagle14-7B base_model: - mlabonne/NeuralBeagle14-7B --- # NeuralBeagle-Flashback NeuralBeagle-Flashback is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing): * [mlabonne/NeuralBeagle14-7B](https://huggingface.co/mlabonne/NeuralBeagle14-7B) ## 🧩 Configuration ```yaml models: - model: timpal0l/Mistral-7B-v0.1-flashback-v2 # No parameters necessary for base model - model: mlabonne/NeuralBeagle14-7B parameters: density: 0.53 weight: 0.6 merge_method: dare_ties base_model: timpal0l/Mistral-7B-v0.1-flashback-v2 parameters: int8_mask: true dtype: bfloat16 ``` ## 💻 Usage ```python !pip install -qU transformers accelerate from transformers import AutoTokenizer import transformers import torch model = "birgermoell/NeuralBeagle-Flashback" messages = [{"role": "user", "content": "What is a large language model?"}] tokenizer = AutoTokenizer.from_pretrained(model) prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) pipeline = transformers.pipeline( "text-generation", model=model, torch_dtype=torch.float16, device_map="auto", ) outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95) print(outputs[0]["generated_text"]) ```
andrewatef/MyBloggerV0.23-main
andrewatef
2024-01-30T19:17:01Z
0
0
transformers
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2024-01-30T19:16:55Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
EzraWilliam/wav2vec2-XLS-R-Fleurs-demo-google-colab-Ezra_William_Prod3
EzraWilliam
2024-01-30T19:14:21Z
3
0
transformers
[ "transformers", "tensorboard", "safetensors", "wav2vec2", "automatic-speech-recognition", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2024-01-29T16:10:07Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
jlbaker361/dcgan-gpu-wikiart25-repeat
jlbaker361
2024-01-30T19:04:29Z
0
0
null
[ "region:us" ]
null
2024-01-30T17:29:26Z
--- {} --- Creative Adversarial Network epochs: 50 dataset jlbaker361/wikiart-balanced25 n classes 27 batch_size 4 images where resized to 768 and then center cropped to: 512 used clip=False discriminator parameters: init_dim: 32 final_dim 512 generator parameters: input noise_dim: 100
OS07/test_code
OS07
2024-01-30T18:58:44Z
63
0
transformers
[ "transformers", "safetensors", "gpt_bigcode", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "bitsandbytes", "region:us" ]
text-generation
2024-01-30T18:58:34Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
TheBloke/CodeLlama-70B-hf-AWQ
TheBloke
2024-01-30T18:55:16Z
14
4
transformers
[ "transformers", "safetensors", "llama", "text-generation", "llama-2", "code", "arxiv:2308.12950", "base_model:codellama/CodeLlama-70b-hf", "base_model:quantized:codellama/CodeLlama-70b-hf", "license:llama2", "autotrain_compatible", "text-generation-inference", "4-bit", "awq", "region:us" ]
text-generation
2024-01-30T16:16:54Z
--- base_model: codellama/CodeLlama-70b-hf inference: false language: - code license: llama2 model_creator: Code Llama model_name: CodeLlama 70B model_type: llama pipeline_tag: text-generation prompt_template: '{prompt} ' quantized_by: TheBloke tags: - llama-2 --- <!-- markdownlint-disable MD041 --> <!-- header start --> <!-- 200823 --> <div style="width: auto; margin-left: auto; margin-right: auto"> <img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;"> </div> <div style="display: flex; justify-content: space-between; width: 100%;"> <div style="display: flex; flex-direction: column; align-items: flex-start;"> <p style="margin-top: 0.5em; margin-bottom: 0em;"><a href="https://discord.gg/theblokeai">Chat & support: TheBloke's Discord server</a></p> </div> <div style="display: flex; flex-direction: column; align-items: flex-end;"> <p style="margin-top: 0.5em; margin-bottom: 0em;"><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? TheBloke's Patreon page</a></p> </div> </div> <div style="text-align:center; margin-top: 0em; margin-bottom: 0em"><p style="margin-top: 0.25em; margin-bottom: 0em;">TheBloke's LLM work is generously supported by a grant from <a href="https://a16z.com">andreessen horowitz (a16z)</a></p></div> <hr style="margin-top: 1.0em; margin-bottom: 1.0em;"> <!-- header end --> # CodeLlama 70B - AWQ - Model creator: [Code Llama](https://huggingface.co/codellama) - Original model: [CodeLlama 70B](https://huggingface.co/codellama/CodeLlama-70b-hf) <!-- description start --> ## Description This repo contains AWQ model files for [Code Llama's CodeLlama 70B](https://huggingface.co/codellama/CodeLlama-70b-hf). These files were quantised using hardware kindly provided by [Massed Compute](https://massedcompute.com/). ### About AWQ AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings. AWQ models are currently supported on Linux and Windows, with NVidia GPUs only. macOS users: please use GGUF models instead. It is supported by: - [Text Generation Webui](https://github.com/oobabooga/text-generation-webui) - using Loader: AutoAWQ - [vLLM](https://github.com/vllm-project/vllm) - version 0.2.2 or later for support for all model types. - [Hugging Face Text Generation Inference (TGI)](https://github.com/huggingface/text-generation-inference) - [Transformers](https://huggingface.co/docs/transformers) version 4.35.0 and later, from any code or client that supports Transformers - [AutoAWQ](https://github.com/casper-hansen/AutoAWQ) - for use from Python code <!-- description end --> <!-- repositories-available start --> ## Repositories available * [AWQ model(s) for GPU inference.](https://huggingface.co/TheBloke/CodeLlama-70B-hf-AWQ) * [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/CodeLlama-70B-hf-GPTQ) * [2, 3, 4, 5, 6 and 8-bit GGUF models for CPU+GPU inference](https://huggingface.co/TheBloke/CodeLlama-70B-hf-GGUF) * [Code Llama's original unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/codellama/CodeLlama-70b-hf) <!-- repositories-available end --> <!-- prompt-template start --> ## Prompt template: None ``` {prompt} ``` <!-- prompt-template end --> <!-- README_AWQ.md-provided-files start --> ## Provided files, and AWQ parameters I currently release 128g GEMM models only. The addition of group_size 32 models, and GEMV kernel models, is being actively considered. Models are released as sharded safetensors files. | Branch | Bits | GS | AWQ Dataset | Seq Len | Size | | ------ | ---- | -- | ----------- | ------- | ---- | | [main](https://huggingface.co/TheBloke/CodeLlama-70B-hf-AWQ/tree/main) | 4 | 128 | [Evol Instruct Code](https://huggingface.co/datasets/nickrosh/Evol-Instruct-Code-80k-v1/viewer/) | 8192 | 36.61 GB <!-- README_AWQ.md-provided-files end --> <!-- README_AWQ.md-text-generation-webui start --> ## How to easily download and use this model in [text-generation-webui](https://github.com/oobabooga/text-generation-webui) Please make sure you're using the latest version of [text-generation-webui](https://github.com/oobabooga/text-generation-webui). It is strongly recommended to use the text-generation-webui one-click-installers unless you're sure you know how to make a manual install. 1. Click the **Model tab**. 2. Under **Download custom model or LoRA**, enter `TheBloke/CodeLlama-70B-hf-AWQ`. 3. Click **Download**. 4. The model will start downloading. Once it's finished it will say "Done". 5. In the top left, click the refresh icon next to **Model**. 6. In the **Model** dropdown, choose the model you just downloaded: `CodeLlama-70B-hf-AWQ` 7. Select **Loader: AutoAWQ**. 8. Click Load, and the model will load and is now ready for use. 9. If you want any custom settings, set them and then click **Save settings for this model** followed by **Reload the Model** in the top right. 10. Once you're ready, click the **Text Generation** tab and enter a prompt to get started! <!-- README_AWQ.md-text-generation-webui end --> <!-- README_AWQ.md-use-from-vllm start --> ## Multi-user inference server: vLLM Documentation on installing and using vLLM [can be found here](https://vllm.readthedocs.io/en/latest/). - Please ensure you are using vLLM version 0.2 or later. - When using vLLM as a server, pass the `--quantization awq` parameter. For example: ```shell python3 -m vllm.entrypoints.api_server --model TheBloke/CodeLlama-70B-hf-AWQ --quantization awq --dtype auto ``` - When using vLLM from Python code, again set `quantization=awq`. For example: ```python from vllm import LLM, SamplingParams prompts = [ "Tell me about AI", "Write a story about llamas", "What is 291 - 150?", "How much wood would a woodchuck chuck if a woodchuck could chuck wood?", ] prompt_template=f'''{prompt} ''' prompts = [prompt_template.format(prompt=prompt) for prompt in prompts] sampling_params = SamplingParams(temperature=0.8, top_p=0.95) llm = LLM(model="TheBloke/CodeLlama-70B-hf-AWQ", quantization="awq", dtype="auto") outputs = llm.generate(prompts, sampling_params) # Print the outputs. for output in outputs: prompt = output.prompt generated_text = output.outputs[0].text print(f"Prompt: {prompt!r}, Generated text: {generated_text!r}") ``` <!-- README_AWQ.md-use-from-vllm start --> <!-- README_AWQ.md-use-from-tgi start --> ## Multi-user inference server: Hugging Face Text Generation Inference (TGI) Use TGI version 1.1.0 or later. The official Docker container is: `ghcr.io/huggingface/text-generation-inference:1.1.0` Example Docker parameters: ```shell --model-id TheBloke/CodeLlama-70B-hf-AWQ --port 3000 --quantize awq --max-input-length 3696 --max-total-tokens 4096 --max-batch-prefill-tokens 4096 ``` Example Python code for interfacing with TGI (requires [huggingface-hub](https://github.com/huggingface/huggingface_hub) 0.17.0 or later): ```shell pip3 install huggingface-hub ``` ```python from huggingface_hub import InferenceClient endpoint_url = "https://your-endpoint-url-here" prompt = "Tell me about AI" prompt_template=f'''{prompt} ''' client = InferenceClient(endpoint_url) response = client.text_generation(prompt, max_new_tokens=128, do_sample=True, temperature=0.7, top_p=0.95, top_k=40, repetition_penalty=1.1) print(f"Model output: ", response) ``` <!-- README_AWQ.md-use-from-tgi end --> <!-- README_AWQ.md-use-from-python start --> ## Inference from Python code using Transformers ### Install the necessary packages - Requires: [Transformers](https://huggingface.co/docs/transformers) 4.35.0 or later. - Requires: [AutoAWQ](https://github.com/casper-hansen/AutoAWQ) 0.1.6 or later. ```shell pip3 install --upgrade "autoawq>=0.1.6" "transformers>=4.35.0" ``` Note that if you are using PyTorch 2.0.1, the above AutoAWQ command will automatically upgrade you to PyTorch 2.1.0. If you are using CUDA 11.8 and wish to continue using PyTorch 2.0.1, instead run this command: ```shell pip3 install https://github.com/casper-hansen/AutoAWQ/releases/download/v0.1.6/autoawq-0.1.6+cu118-cp310-cp310-linux_x86_64.whl ``` If you have problems installing [AutoAWQ](https://github.com/casper-hansen/AutoAWQ) using the pre-built wheels, install it from source instead: ```shell pip3 uninstall -y autoawq git clone https://github.com/casper-hansen/AutoAWQ cd AutoAWQ pip3 install . ``` ### Transformers example code (requires Transformers 4.35.0 and later) ```python from transformers import AutoModelForCausalLM, AutoTokenizer, TextStreamer model_name_or_path = "TheBloke/CodeLlama-70B-hf-AWQ" tokenizer = AutoTokenizer.from_pretrained(model_name_or_path) model = AutoModelForCausalLM.from_pretrained( model_name_or_path, low_cpu_mem_usage=True, device_map="cuda:0" ) # Using the text streamer to stream output one token at a time streamer = TextStreamer(tokenizer, skip_prompt=True, skip_special_tokens=True) prompt = "Tell me about AI" prompt_template=f'''{prompt} ''' # Convert prompt to tokens tokens = tokenizer( prompt_template, return_tensors='pt' ).input_ids.cuda() generation_params = { "do_sample": True, "temperature": 0.7, "top_p": 0.95, "top_k": 40, "max_new_tokens": 512, "repetition_penalty": 1.1 } # Generate streamed output, visible one token at a time generation_output = model.generate( tokens, streamer=streamer, **generation_params ) # Generation without a streamer, which will include the prompt in the output generation_output = model.generate( tokens, **generation_params ) # Get the tokens from the output, decode them, print them token_output = generation_output[0] text_output = tokenizer.decode(token_output) print("model.generate output: ", text_output) # Inference is also possible via Transformers' pipeline from transformers import pipeline pipe = pipeline( "text-generation", model=model, tokenizer=tokenizer, **generation_params ) pipe_output = pipe(prompt_template)[0]['generated_text'] print("pipeline output: ", pipe_output) ``` <!-- README_AWQ.md-use-from-python end --> <!-- README_AWQ.md-compatibility start --> ## Compatibility The files provided are tested to work with: - [text-generation-webui](https://github.com/oobabooga/text-generation-webui) using `Loader: AutoAWQ`. - [vLLM](https://github.com/vllm-project/vllm) version 0.2.0 and later. - [Hugging Face Text Generation Inference (TGI)](https://github.com/huggingface/text-generation-inference) version 1.1.0 and later. - [Transformers](https://huggingface.co/docs/transformers) version 4.35.0 and later. - [AutoAWQ](https://github.com/casper-hansen/AutoAWQ) version 0.1.1 and later. <!-- README_AWQ.md-compatibility end --> <!-- footer start --> <!-- 200823 --> ## Discord For further support, and discussions on these models and AI in general, join us at: [TheBloke AI's Discord server](https://discord.gg/theblokeai) ## Thanks, and how to contribute Thanks to the [chirper.ai](https://chirper.ai) team! Thanks to Clay from [gpus.llm-utils.org](llm-utils)! I've had a lot of people ask if they can contribute. I enjoy providing models and helping people, and would love to be able to spend even more time doing it, as well as expanding into new projects like fine tuning/training. If you're able and willing to contribute it will be most gratefully received and will help me to keep providing more models, and to start work on new AI projects. Donaters will get priority support on any and all AI/LLM/model questions and requests, access to a private Discord room, plus other benefits. * Patreon: https://patreon.com/TheBlokeAI * Ko-Fi: https://ko-fi.com/TheBlokeAI **Special thanks to**: Aemon Algiz. **Patreon special mentions**: Michael Levine, 阿明, Trailburnt, Nikolai Manek, John Detwiler, Randy H, Will Dee, Sebastain Graf, NimbleBox.ai, Eugene Pentland, Emad Mostaque, Ai Maven, Jim Angel, Jeff Scroggin, Michael Davis, Manuel Alberto Morcote, Stephen Murray, Robert, Justin Joy, Luke @flexchar, Brandon Frisco, Elijah Stavena, S_X, Dan Guido, Undi ., Komninos Chatzipapas, Shadi, theTransient, Lone Striker, Raven Klaugh, jjj, Cap'n Zoog, Michel-Marie MAUDET (LINAGORA), Matthew Berman, David, Fen Risland, Omer Bin Jawed, Luke Pendergrass, Kalila, OG, Erik Bjäreholt, Rooh Singh, Joseph William Delisle, Dan Lewis, TL, John Villwock, AzureBlack, Brad, Pedro Madruga, Caitlyn Gatomon, K, jinyuan sun, Mano Prime, Alex, Jeffrey Morgan, Alicia Loh, Illia Dulskyi, Chadd, transmissions 11, fincy, Rainer Wilmers, ReadyPlayerEmma, knownsqashed, Mandus, biorpg, Deo Leter, Brandon Phillips, SuperWojo, Sean Connelly, Iucharbius, Jack West, Harry Royden McLaughlin, Nicholas, terasurfer, Vitor Caleffi, Duane Dunston, Johann-Peter Hartmann, David Ziegler, Olakabola, Ken Nordquist, Trenton Dambrowitz, Tom X Nguyen, Vadim, Ajan Kanaga, Leonard Tan, Clay Pascal, Alexandros Triantafyllidis, JM33133, Xule, vamX, ya boyyy, subjectnull, Talal Aujan, Alps Aficionado, wassieverse, Ari Malik, James Bentley, Woland, Spencer Kim, Michael Dempsey, Fred von Graf, Elle, zynix, William Richards, Stanislav Ovsiannikov, Edmond Seymore, Jonathan Leane, Martin Kemka, usrbinkat, Enrico Ros Thank you to all my generous patrons and donaters! And thank you again to a16z for their generous grant. <!-- footer end --> # Original model card: Code Llama's CodeLlama 70B # **Code Llama** Code Llama is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the repository for the base 70B version in the Hugging Face Transformers format. This model is designed for general code synthesis and understanding. Links to other models can be found in the index at the bottom. | | Base Model | Python | Instruct | | --- | ----------------------------------------------------------------------------- | ------------------------------------------------------------------------------------------- | ----------------------------------------------------------------------------------------------- | | 7B | [codellama/CodeLlama-7b-hf](https://huggingface.co/codellama/CodeLlama-7b-hf) | [codellama/CodeLlama-7b-Python-hf](https://huggingface.co/codellama/CodeLlama-7b-Python-hf) | [codellama/CodeLlama-7b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-7b-Instruct-hf) | | 13B | [codellama/CodeLlama-13b-hf](https://huggingface.co/codellama/CodeLlama-13b-hf) | [codellama/CodeLlama-13b-Python-hf](https://huggingface.co/codellama/CodeLlama-13b-Python-hf) | [codellama/CodeLlama-13b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-13b-Instruct-hf) | | 34B | [codellama/CodeLlama-34b-hf](https://huggingface.co/codellama/CodeLlama-34b-hf) | [codellama/CodeLlama-34b-Python-hf](https://huggingface.co/codellama/CodeLlama-34b-Python-hf) | [codellama/CodeLlama-34b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-34b-Instruct-hf) | | 70B | [codellama/CodeLlama-70b-hf](https://huggingface.co/codellama/CodeLlama-70b-hf) | [codellama/CodeLlama-70b-Python-hf](https://huggingface.co/codellama/CodeLlama-70b-Python-hf) | [codellama/CodeLlama-70b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-70b-Instruct-hf) | ## Model Use To use this model, please make sure to install `transformers`. ```bash pip install transformers accelerate ``` Model capabilities: - [x] Code completion. - [ ] Infilling. - [ ] Instructions / chat. - [ ] Python specialist. ## Model Details *Note: Use of this model is governed by the Meta license. Meta developed and publicly released the Code Llama family of large language models (LLMs). **Model Developers** Meta **Variations** Code Llama comes in four model sizes, and three variants: * Code Llama: base models designed for general code synthesis and understanding * Code Llama - Python: designed specifically for Python * Code Llama - Instruct: for instruction following and safer deployment All variants are available in sizes of 7B, 13B, 34B, and 70B parameters. **This repository contains the base version of the 70B parameters model.** **Input** Models input text only. **Output** Models generate text only. **Model Architecture** Code Llama is an auto-regressive language model that uses an optimized transformer architecture. It was fine-tuned with up to 16k tokens and supports up to 100k tokens at inference time. **Model Dates** Code Llama and its variants have been trained between January 2023 and January 2024. **Status** This is a static model trained on an offline dataset. Future versions of Code Llama - Instruct will be released as we improve model safety with community feedback. **License** A custom commercial license is available at: [https://ai.meta.com/resources/models-and-libraries/llama-downloads/](https://ai.meta.com/resources/models-and-libraries/llama-downloads/) **Research Paper** More information can be found in the paper "[Code Llama: Open Foundation Models for Code](https://ai.meta.com/research/publications/code-llama-open-foundation-models-for-code/)" or its [arXiv page](https://arxiv.org/abs/2308.12950). ## Intended Use **Intended Use Cases** Code Llama and its variants are intended for commercial and research use in English and relevant programming languages. The base model Code Llama can be adapted for a variety of code synthesis and understanding tasks, Code Llama - Python is designed specifically to handle the Python programming language, and Code Llama - Instruct is intended to be safer to use for code assistant and generation applications. **Out-of-Scope Uses** Use in any manner that violates applicable laws or regulations (including trade compliance laws). Use in languages other than English. Use in any other way that is prohibited by the Acceptable Use Policy and Licensing Agreement for Code Llama and its variants. ## Hardware and Software **Training Factors** We used custom training libraries. The training and fine-tuning of the released models have been performed Meta’s Research Super Cluster. **Carbon Footprint** In aggregate, training all 12 Code Llama models required 1400K GPU hours of computation on hardware of type A100-80GB (TDP of 350-400W). Estimated total emissions were 228.55 tCO2eq, 100% of which were offset by Meta’s sustainability program. ## Evaluation Results See evaluations for the main models and detailed ablations in Section 3 and safety evaluations in Section 4 of the research paper. ## Ethical Considerations and Limitations Code Llama and its variants are a new technology that carries risks with use. Testing conducted to date has been in English, and has not covered, nor could it cover all scenarios. For these reasons, as with all LLMs, Code Llama’s potential outputs cannot be predicted in advance, and the model may in some instances produce inaccurate or objectionable responses to user prompts. Therefore, before deploying any applications of Code Llama, developers should perform safety testing and tuning tailored to their specific applications of the model. Please see the Responsible Use Guide available available at [https://ai.meta.com/llama/responsible-use-guide](https://ai.meta.com/llama/responsible-use-guide).
espirado1/emotion-test-1000
espirado1
2024-01-30T18:45:13Z
92
2
transformers
[ "transformers", "safetensors", "distilbert", "text-classification", "intel", "ipex", "bf16", "en", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-01-30T18:30:16Z
--- license: mit language: - en tags: - intel - ipex - bf16 --- --- # Model Card for My Fine-Tuned Model ## Model Description - **Purpose**: This model is fine-tuned to perform multi-class emotion classification. It can identify various emotions in text, such as joy, sadness, love, anger, fear, and surprise. - **Model architecture**: The model is based on the distilbert-base-uncased architecture, a distilled version of the BERT model which is smaller and faster but retains most of its predictive power - **Training data**: The model was trained on the emotion dataset from Hugging Face's datasets library. This dataset includes text labeled with different emotions. During preprocessing, texts were tokenized, and padding and truncation were applied to standardize their lengths ## Intended Use - **Intended users**: This model is intended for developers and researchers interested in emotion analysis in text, including applications in social media sentiment analysis, customer feedback interpretation, and mental health assessment. - **Use cases**: Potential use cases include analyzing social media posts for emotional content, enhancing chatbots to understand user emotions, and helping mental health professionals in identifying emotional states from text-based communications. ## Limitations - **Known limitations**: The model's accuracy may vary depending on the context and the dataset's representativeness. It may not perform equally well on texts from domains significantly different from the training data. ## Hardware - **Training Platform**: The model was trained on 4th Generation Intel Xeon Processors available on the Intel Developer Cloud (cloud.intel.com). The training completed in under 8 minutes, demonstrating the efficiency of Intel hardware optimizations. ## Ethical Considerations - **Ethical concerns**: Care should be taken to ensure that the model is not used in sensitive applications without proper ethical considerations, especially in scenarios that could impact individual privacy or mental health. ## More Information ### Software Optimizations - **Known Optimizations**: Training Setup: The training leveraged Intel extensions for PyTorch (IPEX) to optimize training efficiency on Intel hardware. Mixed precision training (FP32 and BF16) was enabled, contributing to the rapid training time.
eagle0504/eagle0504-mistral-7b-v0.2
eagle0504
2024-01-30T18:44:36Z
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:mistralai/Mistral-7B-v0.1", "base_model:adapter:mistralai/Mistral-7B-v0.1", "region:us" ]
null
2024-01-30T18:39:26Z
--- library_name: peft base_model: mistralai/Mistral-7B-v0.1 --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.8.1
philimon/tinyllama-esg-lora
philimon
2024-01-30T18:43:39Z
0
0
null
[ "tensorboard", "safetensors", "trl", "sft", "generated_from_trainer", "base_model:TinyLlama/TinyLlama-1.1B-Chat-v0.3", "base_model:finetune:TinyLlama/TinyLlama-1.1B-Chat-v0.3", "license:apache-2.0", "region:us" ]
null
2024-01-30T18:21:14Z
--- license: apache-2.0 base_model: PY007/TinyLlama-1.1B-Chat-v0.3 tags: - trl - sft - generated_from_trainer model-index: - name: tinyllama-esg-lora results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # tinyllama-esg-lora This model is a fine-tuned version of [PY007/TinyLlama-1.1B-Chat-v0.3](https://huggingface.co/PY007/TinyLlama-1.1B-Chat-v0.3) on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 32 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - training_steps: 500 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.16.1 - Tokenizers 0.15.1
gjonesQ02/testLLMModels
gjonesQ02
2024-01-30T18:40:05Z
109
0
transformers
[ "transformers", "tensorboard", "safetensors", "distilbert", "question-answering", "generated_from_trainer", "base_model:distilbert/distilbert-base-uncased", "base_model:finetune:distilbert/distilbert-base-uncased", "license:apache-2.0", "endpoints_compatible", "region:us" ]
question-answering
2023-12-18T20:59:32Z
--- license: apache-2.0 base_model: distilbert-base-uncased tags: - generated_from_trainer model-index: - name: testLLMModels results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # testLLMModels This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 1.7927 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | No log | 1.0 | 250 | 2.3645 | | 2.7423 | 2.0 | 500 | 1.8809 | | 2.7423 | 3.0 | 750 | 1.7927 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.1.0+cu121 - Datasets 2.16.1 - Tokenizers 0.15.1
superamar/peacock
superamar
2024-01-30T18:33:10Z
3
0
diffusers
[ "diffusers", "safetensors", "NxtWave-GenAI-Webinar", "text-to-image", "stable-diffusion", "license:creativeml-openrail-m", "autotrain_compatible", "endpoints_compatible", "diffusers:StableDiffusionPipeline", "region:us" ]
text-to-image
2024-01-30T18:28:33Z
--- license: creativeml-openrail-m tags: - NxtWave-GenAI-Webinar - text-to-image - stable-diffusion --- ### Peacock Dreambooth model trained by superamar following the "Build your own Gen AI model" session by NxtWave. Project Submission Code: 38732523004 Sample pictures of this concept: ![0](https://huggingface.co/superamar/peacock/resolve/main/sample_images/40d10efd-d673-477a-ab82-e745319dd70c.jpeg)
floriangardin/chord_model
floriangardin
2024-01-30T18:17:03Z
187
0
transformers
[ "transformers", "tensorboard", "safetensors", "gpt2", "text-generation", "generated_from_trainer", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-01-30T17:12:00Z
--- tags: - generated_from_trainer model-index: - name: chord_model results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # chord_model This model is a fine-tuned version of [](https://huggingface.co/) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.4598 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 8 - eval_batch_size: 8 - seed: 444 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine_with_restarts - lr_scheduler_warmup_ratio: 0.3 - training_steps: 0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:-----:|:---------------:| | 1.2264 | 0.11 | 500 | 1.1269 | | 0.9624 | 0.21 | 1000 | 0.9066 | | 0.8598 | 0.32 | 1500 | 0.8128 | | 0.8209 | 0.43 | 2000 | 0.7626 | | 0.7483 | 0.53 | 2500 | 0.7272 | | 0.7391 | 0.64 | 3000 | 0.7032 | | 0.7052 | 0.75 | 3500 | 0.6739 | | 0.6998 | 0.86 | 4000 | 0.6503 | | 0.6901 | 0.96 | 4500 | 0.6244 | | 0.6348 | 1.07 | 5000 | 0.6100 | | 0.654 | 1.18 | 5500 | 0.5891 | | 0.6227 | 1.28 | 6000 | 0.5765 | | 0.6148 | 1.39 | 6500 | 0.5624 | | 0.5973 | 1.5 | 7000 | 0.5538 | | 0.5853 | 1.6 | 7500 | 0.5441 | | 0.56 | 1.71 | 8000 | 0.5407 | | 0.574 | 1.82 | 8500 | 0.5342 | | 0.5589 | 1.92 | 9000 | 0.5296 | | 0.5634 | 2.03 | 9500 | 0.5254 | | 0.543 | 2.14 | 10000 | 0.5208 | | 0.5792 | 2.25 | 10500 | 0.5159 | | 0.5571 | 2.35 | 11000 | 0.5064 | | 0.5408 | 2.46 | 11500 | 0.4957 | | 0.5398 | 2.57 | 12000 | 0.4882 | | 0.537 | 2.67 | 12500 | 0.4834 | | 0.5512 | 2.78 | 13000 | 0.4786 | | 0.4842 | 2.89 | 13500 | 0.4753 | | 0.5275 | 2.99 | 14000 | 0.4721 | | 0.4899 | 3.1 | 14500 | 0.4710 | | 0.5222 | 3.21 | 15000 | 0.4666 | | 0.4929 | 3.31 | 15500 | 0.4645 | | 0.5049 | 3.42 | 16000 | 0.4631 | | 0.5002 | 3.53 | 16500 | 0.4613 | | 0.505 | 3.64 | 17000 | 0.4611 | | 0.507 | 3.74 | 17500 | 0.4602 | | 0.5169 | 3.85 | 18000 | 0.4598 | | 0.501 | 3.96 | 18500 | 0.4598 | ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.16.1 - Tokenizers 0.15.1
TheBloke/CodeLlama-70B-hf-GGUF
TheBloke
2024-01-30T18:13:10Z
773
43
transformers
[ "transformers", "gguf", "llama", "llama-2", "text-generation", "code", "arxiv:2308.12950", "base_model:codellama/CodeLlama-70b-hf", "base_model:quantized:codellama/CodeLlama-70b-hf", "license:llama2", "region:us" ]
text-generation
2024-01-30T16:16:54Z
--- base_model: codellama/CodeLlama-70b-hf inference: false language: - code license: llama2 model_creator: Code Llama model_name: CodeLlama 70B model_type: llama pipeline_tag: text-generation prompt_template: '{prompt} ' quantized_by: TheBloke tags: - llama-2 --- <!-- markdownlint-disable MD041 --> <!-- header start --> <!-- 200823 --> <div style="width: auto; margin-left: auto; margin-right: auto"> <img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;"> </div> <div style="display: flex; justify-content: space-between; width: 100%;"> <div style="display: flex; flex-direction: column; align-items: flex-start;"> <p style="margin-top: 0.5em; margin-bottom: 0em;"><a href="https://discord.gg/theblokeai">Chat & support: TheBloke's Discord server</a></p> </div> <div style="display: flex; flex-direction: column; align-items: flex-end;"> <p style="margin-top: 0.5em; margin-bottom: 0em;"><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? TheBloke's Patreon page</a></p> </div> </div> <div style="text-align:center; margin-top: 0em; margin-bottom: 0em"><p style="margin-top: 0.25em; margin-bottom: 0em;">TheBloke's LLM work is generously supported by a grant from <a href="https://a16z.com">andreessen horowitz (a16z)</a></p></div> <hr style="margin-top: 1.0em; margin-bottom: 1.0em;"> <!-- header end --> # CodeLlama 70B - GGUF - Model creator: [Code Llama](https://huggingface.co/codellama) - Original model: [CodeLlama 70B](https://huggingface.co/codellama/CodeLlama-70b-hf) <!-- description start --> ## Description This repo contains GGUF format model files for [Code Llama's CodeLlama 70B](https://huggingface.co/codellama/CodeLlama-70b-hf). These files were quantised using hardware kindly provided by [Massed Compute](https://massedcompute.com/). <!-- description end --> <!-- README_GGUF.md-about-gguf start --> ### About GGUF GGUF is a new format introduced by the llama.cpp team on August 21st 2023. It is a replacement for GGML, which is no longer supported by llama.cpp. Here is an incomplete list of clients and libraries that are known to support GGUF: * [llama.cpp](https://github.com/ggerganov/llama.cpp). The source project for GGUF. Offers a CLI and a server option. * [text-generation-webui](https://github.com/oobabooga/text-generation-webui), the most widely used web UI, with many features and powerful extensions. Supports GPU acceleration. * [KoboldCpp](https://github.com/LostRuins/koboldcpp), a fully featured web UI, with GPU accel across all platforms and GPU architectures. Especially good for story telling. * [GPT4All](https://gpt4all.io/index.html), a free and open source local running GUI, supporting Windows, Linux and macOS with full GPU accel. * [LM Studio](https://lmstudio.ai/), an easy-to-use and powerful local GUI for Windows and macOS (Silicon), with GPU acceleration. Linux available, in beta as of 27/11/2023. * [LoLLMS Web UI](https://github.com/ParisNeo/lollms-webui), a great web UI with many interesting and unique features, including a full model library for easy model selection. * [Faraday.dev](https://faraday.dev/), an attractive and easy to use character-based chat GUI for Windows and macOS (both Silicon and Intel), with GPU acceleration. * [llama-cpp-python](https://github.com/abetlen/llama-cpp-python), a Python library with GPU accel, LangChain support, and OpenAI-compatible API server. * [candle](https://github.com/huggingface/candle), a Rust ML framework with a focus on performance, including GPU support, and ease of use. * [ctransformers](https://github.com/marella/ctransformers), a Python library with GPU accel, LangChain support, and OpenAI-compatible AI server. Note, as of time of writing (November 27th 2023), ctransformers has not been updated in a long time and does not support many recent models. <!-- README_GGUF.md-about-gguf end --> <!-- repositories-available start --> ## Repositories available * [AWQ model(s) for GPU inference.](https://huggingface.co/TheBloke/CodeLlama-70B-hf-AWQ) * [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/CodeLlama-70B-hf-GPTQ) * [2, 3, 4, 5, 6 and 8-bit GGUF models for CPU+GPU inference](https://huggingface.co/TheBloke/CodeLlama-70B-hf-GGUF) * [Code Llama's original unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/codellama/CodeLlama-70b-hf) <!-- repositories-available end --> <!-- prompt-template start --> ## Prompt template: None ``` {prompt} ``` <!-- prompt-template end --> <!-- compatibility_gguf start --> ## Compatibility These quantised GGUFv2 files are compatible with llama.cpp from August 27th onwards, as of commit [d0cee0d](https://github.com/ggerganov/llama.cpp/commit/d0cee0d36d5be95a0d9088b674dbb27354107221) They are also compatible with many third party UIs and libraries - please see the list at the top of this README. ## Explanation of quantisation methods <details> <summary>Click to see details</summary> The new methods available are: * GGML_TYPE_Q2_K - "type-1" 2-bit quantization in super-blocks containing 16 blocks, each block having 16 weight. Block scales and mins are quantized with 4 bits. This ends up effectively using 2.5625 bits per weight (bpw) * GGML_TYPE_Q3_K - "type-0" 3-bit quantization in super-blocks containing 16 blocks, each block having 16 weights. Scales are quantized with 6 bits. This end up using 3.4375 bpw. * GGML_TYPE_Q4_K - "type-1" 4-bit quantization in super-blocks containing 8 blocks, each block having 32 weights. Scales and mins are quantized with 6 bits. This ends up using 4.5 bpw. * GGML_TYPE_Q5_K - "type-1" 5-bit quantization. Same super-block structure as GGML_TYPE_Q4_K resulting in 5.5 bpw * GGML_TYPE_Q6_K - "type-0" 6-bit quantization. Super-blocks with 16 blocks, each block having 16 weights. Scales are quantized with 8 bits. This ends up using 6.5625 bpw Refer to the Provided Files table below to see what files use which methods, and how. </details> <!-- compatibility_gguf end --> <!-- README_GGUF.md-provided-files start --> ## Provided files | Name | Quant method | Bits | Size | Max RAM required | Use case | | ---- | ---- | ---- | ---- | ---- | ----- | | [codellama-70b-hf.Q2_K.gguf](https://huggingface.co/TheBloke/CodeLlama-70B-hf-GGUF/blob/main/codellama-70b-hf.Q2_K.gguf) | Q2_K | 2 | 25.46 GB| 27.96 GB | significant quality loss - not recommended for most purposes | | [codellama-70b-hf.Q3_K_S.gguf](https://huggingface.co/TheBloke/CodeLlama-70B-hf-GGUF/blob/main/codellama-70b-hf.Q3_K_S.gguf) | Q3_K_S | 3 | 29.92 GB| 32.42 GB | very small, high quality loss | | [codellama-70b-hf.Q3_K_M.gguf](https://huggingface.co/TheBloke/CodeLlama-70B-hf-GGUF/blob/main/codellama-70b-hf.Q3_K_M.gguf) | Q3_K_M | 3 | 33.27 GB| 35.77 GB | very small, high quality loss | | [codellama-70b-hf.Q3_K_L.gguf](https://huggingface.co/TheBloke/CodeLlama-70B-hf-GGUF/blob/main/codellama-70b-hf.Q3_K_L.gguf) | Q3_K_L | 3 | 36.15 GB| 38.65 GB | small, substantial quality loss | | [codellama-70b-hf.Q4_0.gguf](https://huggingface.co/TheBloke/CodeLlama-70B-hf-GGUF/blob/main/codellama-70b-hf.Q4_0.gguf) | Q4_0 | 4 | 38.87 GB| 41.37 GB | legacy; small, very high quality loss - prefer using Q3_K_M | | [codellama-70b-hf.Q4_K_S.gguf](https://huggingface.co/TheBloke/CodeLlama-70B-hf-GGUF/blob/main/codellama-70b-hf.Q4_K_S.gguf) | Q4_K_S | 4 | 39.25 GB| 41.75 GB | small, greater quality loss | | [codellama-70b-hf.Q4_K_M.gguf](https://huggingface.co/TheBloke/CodeLlama-70B-hf-GGUF/blob/main/codellama-70b-hf.Q4_K_M.gguf) | Q4_K_M | 4 | 41.42 GB| 43.92 GB | medium, balanced quality - recommended | | [codellama-70b-hf.Q5_0.gguf](https://huggingface.co/TheBloke/CodeLlama-70B-hf-GGUF/blob/main/codellama-70b-hf.Q5_0.gguf) | Q5_0 | 5 | 47.46 GB| 49.96 GB | legacy; medium, balanced quality - prefer using Q4_K_M | | [codellama-70b-hf.Q5_K_S.gguf](https://huggingface.co/TheBloke/CodeLlama-70B-hf-GGUF/blob/main/codellama-70b-hf.Q5_K_S.gguf) | Q5_K_S | 5 | 47.46 GB| 49.96 GB | large, low quality loss - recommended | | [codellama-70b-hf.Q5_K_M.gguf](https://huggingface.co/TheBloke/CodeLlama-70B-hf-GGUF/blob/main/codellama-70b-hf.Q5_K_M.gguf) | Q5_K_M | 5 | 48.75 GB| 51.25 GB | large, very low quality loss - recommended | | codellama-70b-hf.Q6_K.gguf | Q6_K | 6 | 56.59 GB| 59.09 GB | very large, extremely low quality loss | | codellama-70b-hf.Q8_0.gguf | Q8_0 | 8 | 73.29 GB| 75.79 GB | very large, extremely low quality loss - not recommended | **Note**: the above RAM figures assume no GPU offloading. If layers are offloaded to the GPU, this will reduce RAM usage and use VRAM instead. ### Q6_K and Q8_0 files are split and require joining **Note:** HF does not support uploading files larger than 50GB. Therefore I have uploaded the Q6_K and Q8_0 files as split files. <details> <summary>Click for instructions regarding Q6_K and Q8_0 files</summary> ### q6_K Please download: * `codellama-70b-hf.Q6_K.gguf-split-a` * `codellama-70b-hf.Q6_K.gguf-split-b` ### q8_0 Please download: * `codellama-70b-hf.Q8_0.gguf-split-a` * `codellama-70b-hf.Q8_0.gguf-split-b` To join the files, do the following: Linux and macOS: ``` cat codellama-70b-hf.Q6_K.gguf-split-* > codellama-70b-hf.Q6_K.gguf && rm codellama-70b-hf.Q6_K.gguf-split-* cat codellama-70b-hf.Q8_0.gguf-split-* > codellama-70b-hf.Q8_0.gguf && rm codellama-70b-hf.Q8_0.gguf-split-* ``` Windows command line: ``` COPY /B codellama-70b-hf.Q6_K.gguf-split-a + codellama-70b-hf.Q6_K.gguf-split-b codellama-70b-hf.Q6_K.gguf del codellama-70b-hf.Q6_K.gguf-split-a codellama-70b-hf.Q6_K.gguf-split-b COPY /B codellama-70b-hf.Q8_0.gguf-split-a + codellama-70b-hf.Q8_0.gguf-split-b codellama-70b-hf.Q8_0.gguf del codellama-70b-hf.Q8_0.gguf-split-a codellama-70b-hf.Q8_0.gguf-split-b ``` </details> <!-- README_GGUF.md-provided-files end --> <!-- README_GGUF.md-how-to-download start --> ## How to download GGUF files **Note for manual downloaders:** You almost never want to clone the entire repo! Multiple different quantisation formats are provided, and most users only want to pick and download a single file. The following clients/libraries will automatically download models for you, providing a list of available models to choose from: * LM Studio * LoLLMS Web UI * Faraday.dev ### In `text-generation-webui` Under Download Model, you can enter the model repo: TheBloke/CodeLlama-70B-hf-GGUF and below it, a specific filename to download, such as: codellama-70b-hf.Q4_K_M.gguf. Then click Download. ### On the command line, including multiple files at once I recommend using the `huggingface-hub` Python library: ```shell pip3 install huggingface-hub ``` Then you can download any individual model file to the current directory, at high speed, with a command like this: ```shell huggingface-cli download TheBloke/CodeLlama-70B-hf-GGUF codellama-70b-hf.Q4_K_M.gguf --local-dir . --local-dir-use-symlinks False ``` <details> <summary>More advanced huggingface-cli download usage (click to read)</summary> You can also download multiple files at once with a pattern: ```shell huggingface-cli download TheBloke/CodeLlama-70B-hf-GGUF --local-dir . --local-dir-use-symlinks False --include='*Q4_K*gguf' ``` For more documentation on downloading with `huggingface-cli`, please see: [HF -> Hub Python Library -> Download files -> Download from the CLI](https://huggingface.co/docs/huggingface_hub/guides/download#download-from-the-cli). To accelerate downloads on fast connections (1Gbit/s or higher), install `hf_transfer`: ```shell pip3 install hf_transfer ``` And set environment variable `HF_HUB_ENABLE_HF_TRANSFER` to `1`: ```shell HF_HUB_ENABLE_HF_TRANSFER=1 huggingface-cli download TheBloke/CodeLlama-70B-hf-GGUF codellama-70b-hf.Q4_K_M.gguf --local-dir . --local-dir-use-symlinks False ``` Windows Command Line users: You can set the environment variable by running `set HF_HUB_ENABLE_HF_TRANSFER=1` before the download command. </details> <!-- README_GGUF.md-how-to-download end --> <!-- README_GGUF.md-how-to-run start --> ## Example `llama.cpp` command Make sure you are using `llama.cpp` from commit [d0cee0d](https://github.com/ggerganov/llama.cpp/commit/d0cee0d36d5be95a0d9088b674dbb27354107221) or later. ```shell ./main -ngl 35 -m codellama-70b-hf.Q4_K_M.gguf --color -c 16384 --temp 0.7 --repeat_penalty 1.1 -n -1 -p "{prompt}" ``` Change `-ngl 32` to the number of layers to offload to GPU. Remove it if you don't have GPU acceleration. Change `-c 16384` to the desired sequence length. For extended sequence models - eg 8K, 16K, 32K - the necessary RoPE scaling parameters are read from the GGUF file and set by llama.cpp automatically. Note that longer sequence lengths require much more resources, so you may need to reduce this value. If you want to have a chat-style conversation, replace the `-p <PROMPT>` argument with `-i -ins` For other parameters and how to use them, please refer to [the llama.cpp documentation](https://github.com/ggerganov/llama.cpp/blob/master/examples/main/README.md) ## How to run in `text-generation-webui` Further instructions can be found in the text-generation-webui documentation, here: [text-generation-webui/docs/04 ‐ Model Tab.md](https://github.com/oobabooga/text-generation-webui/blob/main/docs/04%20%E2%80%90%20Model%20Tab.md#llamacpp). ## How to run from Python code You can use GGUF models from Python using the [llama-cpp-python](https://github.com/abetlen/llama-cpp-python) or [ctransformers](https://github.com/marella/ctransformers) libraries. Note that at the time of writing (Nov 27th 2023), ctransformers has not been updated for some time and is not compatible with some recent models. Therefore I recommend you use llama-cpp-python. ### How to load this model in Python code, using llama-cpp-python For full documentation, please see: [llama-cpp-python docs](https://abetlen.github.io/llama-cpp-python/). #### First install the package Run one of the following commands, according to your system: ```shell # Base ctransformers with no GPU acceleration pip install llama-cpp-python # With NVidia CUDA acceleration CMAKE_ARGS="-DLLAMA_CUBLAS=on" pip install llama-cpp-python # Or with OpenBLAS acceleration CMAKE_ARGS="-DLLAMA_BLAS=ON -DLLAMA_BLAS_VENDOR=OpenBLAS" pip install llama-cpp-python # Or with CLBLast acceleration CMAKE_ARGS="-DLLAMA_CLBLAST=on" pip install llama-cpp-python # Or with AMD ROCm GPU acceleration (Linux only) CMAKE_ARGS="-DLLAMA_HIPBLAS=on" pip install llama-cpp-python # Or with Metal GPU acceleration for macOS systems only CMAKE_ARGS="-DLLAMA_METAL=on" pip install llama-cpp-python # In windows, to set the variables CMAKE_ARGS in PowerShell, follow this format; eg for NVidia CUDA: $env:CMAKE_ARGS = "-DLLAMA_OPENBLAS=on" pip install llama-cpp-python ``` #### Simple llama-cpp-python example code ```python from llama_cpp import Llama # Set gpu_layers to the number of layers to offload to GPU. Set to 0 if no GPU acceleration is available on your system. llm = Llama( model_path="./codellama-70b-hf.Q4_K_M.gguf", # Download the model file first n_ctx=16384, # The max sequence length to use - note that longer sequence lengths require much more resources n_threads=8, # The number of CPU threads to use, tailor to your system and the resulting performance n_gpu_layers=35 # The number of layers to offload to GPU, if you have GPU acceleration available ) # Simple inference example output = llm( "{prompt}", # Prompt max_tokens=512, # Generate up to 512 tokens stop=["</s>"], # Example stop token - not necessarily correct for this specific model! Please check before using. echo=True # Whether to echo the prompt ) # Chat Completion API llm = Llama(model_path="./codellama-70b-hf.Q4_K_M.gguf", chat_format="llama-2") # Set chat_format according to the model you are using llm.create_chat_completion( messages = [ {"role": "system", "content": "You are a story writing assistant."}, { "role": "user", "content": "Write a story about llamas." } ] ) ``` ## How to use with LangChain Here are guides on using llama-cpp-python and ctransformers with LangChain: * [LangChain + llama-cpp-python](https://python.langchain.com/docs/integrations/llms/llamacpp) * [LangChain + ctransformers](https://python.langchain.com/docs/integrations/providers/ctransformers) <!-- README_GGUF.md-how-to-run end --> <!-- footer start --> <!-- 200823 --> ## Discord For further support, and discussions on these models and AI in general, join us at: [TheBloke AI's Discord server](https://discord.gg/theblokeai) ## Thanks, and how to contribute Thanks to the [chirper.ai](https://chirper.ai) team! Thanks to Clay from [gpus.llm-utils.org](llm-utils)! I've had a lot of people ask if they can contribute. I enjoy providing models and helping people, and would love to be able to spend even more time doing it, as well as expanding into new projects like fine tuning/training. If you're able and willing to contribute it will be most gratefully received and will help me to keep providing more models, and to start work on new AI projects. Donaters will get priority support on any and all AI/LLM/model questions and requests, access to a private Discord room, plus other benefits. * Patreon: https://patreon.com/TheBlokeAI * Ko-Fi: https://ko-fi.com/TheBlokeAI **Special thanks to**: Aemon Algiz. **Patreon special mentions**: Michael Levine, 阿明, Trailburnt, Nikolai Manek, John Detwiler, Randy H, Will Dee, Sebastain Graf, NimbleBox.ai, Eugene Pentland, Emad Mostaque, Ai Maven, Jim Angel, Jeff Scroggin, Michael Davis, Manuel Alberto Morcote, Stephen Murray, Robert, Justin Joy, Luke @flexchar, Brandon Frisco, Elijah Stavena, S_X, Dan Guido, Undi ., Komninos Chatzipapas, Shadi, theTransient, Lone Striker, Raven Klaugh, jjj, Cap'n Zoog, Michel-Marie MAUDET (LINAGORA), Matthew Berman, David, Fen Risland, Omer Bin Jawed, Luke Pendergrass, Kalila, OG, Erik Bjäreholt, Rooh Singh, Joseph William Delisle, Dan Lewis, TL, John Villwock, AzureBlack, Brad, Pedro Madruga, Caitlyn Gatomon, K, jinyuan sun, Mano Prime, Alex, Jeffrey Morgan, Alicia Loh, Illia Dulskyi, Chadd, transmissions 11, fincy, Rainer Wilmers, ReadyPlayerEmma, knownsqashed, Mandus, biorpg, Deo Leter, Brandon Phillips, SuperWojo, Sean Connelly, Iucharbius, Jack West, Harry Royden McLaughlin, Nicholas, terasurfer, Vitor Caleffi, Duane Dunston, Johann-Peter Hartmann, David Ziegler, Olakabola, Ken Nordquist, Trenton Dambrowitz, Tom X Nguyen, Vadim, Ajan Kanaga, Leonard Tan, Clay Pascal, Alexandros Triantafyllidis, JM33133, Xule, vamX, ya boyyy, subjectnull, Talal Aujan, Alps Aficionado, wassieverse, Ari Malik, James Bentley, Woland, Spencer Kim, Michael Dempsey, Fred von Graf, Elle, zynix, William Richards, Stanislav Ovsiannikov, Edmond Seymore, Jonathan Leane, Martin Kemka, usrbinkat, Enrico Ros Thank you to all my generous patrons and donaters! And thank you again to a16z for their generous grant. <!-- footer end --> <!-- original-model-card start --> # Original model card: Code Llama's CodeLlama 70B # **Code Llama** Code Llama is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the repository for the base 70B version in the Hugging Face Transformers format. This model is designed for general code synthesis and understanding. Links to other models can be found in the index at the bottom. | | Base Model | Python | Instruct | | --- | ----------------------------------------------------------------------------- | ------------------------------------------------------------------------------------------- | ----------------------------------------------------------------------------------------------- | | 7B | [codellama/CodeLlama-7b-hf](https://huggingface.co/codellama/CodeLlama-7b-hf) | [codellama/CodeLlama-7b-Python-hf](https://huggingface.co/codellama/CodeLlama-7b-Python-hf) | [codellama/CodeLlama-7b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-7b-Instruct-hf) | | 13B | [codellama/CodeLlama-13b-hf](https://huggingface.co/codellama/CodeLlama-13b-hf) | [codellama/CodeLlama-13b-Python-hf](https://huggingface.co/codellama/CodeLlama-13b-Python-hf) | [codellama/CodeLlama-13b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-13b-Instruct-hf) | | 34B | [codellama/CodeLlama-34b-hf](https://huggingface.co/codellama/CodeLlama-34b-hf) | [codellama/CodeLlama-34b-Python-hf](https://huggingface.co/codellama/CodeLlama-34b-Python-hf) | [codellama/CodeLlama-34b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-34b-Instruct-hf) | | 70B | [codellama/CodeLlama-70b-hf](https://huggingface.co/codellama/CodeLlama-70b-hf) | [codellama/CodeLlama-70b-Python-hf](https://huggingface.co/codellama/CodeLlama-70b-Python-hf) | [codellama/CodeLlama-70b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-70b-Instruct-hf) | ## Model Use To use this model, please make sure to install `transformers`. ```bash pip install transformers accelerate ``` Model capabilities: - [x] Code completion. - [ ] Infilling. - [ ] Instructions / chat. - [ ] Python specialist. ## Model Details *Note: Use of this model is governed by the Meta license. Meta developed and publicly released the Code Llama family of large language models (LLMs). **Model Developers** Meta **Variations** Code Llama comes in four model sizes, and three variants: * Code Llama: base models designed for general code synthesis and understanding * Code Llama - Python: designed specifically for Python * Code Llama - Instruct: for instruction following and safer deployment All variants are available in sizes of 7B, 13B, 34B, and 70B parameters. **This repository contains the base version of the 70B parameters model.** **Input** Models input text only. **Output** Models generate text only. **Model Architecture** Code Llama is an auto-regressive language model that uses an optimized transformer architecture. It was fine-tuned with up to 16k tokens and supports up to 100k tokens at inference time. **Model Dates** Code Llama and its variants have been trained between January 2023 and January 2024. **Status** This is a static model trained on an offline dataset. Future versions of Code Llama - Instruct will be released as we improve model safety with community feedback. **License** A custom commercial license is available at: [https://ai.meta.com/resources/models-and-libraries/llama-downloads/](https://ai.meta.com/resources/models-and-libraries/llama-downloads/) **Research Paper** More information can be found in the paper "[Code Llama: Open Foundation Models for Code](https://ai.meta.com/research/publications/code-llama-open-foundation-models-for-code/)" or its [arXiv page](https://arxiv.org/abs/2308.12950). ## Intended Use **Intended Use Cases** Code Llama and its variants are intended for commercial and research use in English and relevant programming languages. The base model Code Llama can be adapted for a variety of code synthesis and understanding tasks, Code Llama - Python is designed specifically to handle the Python programming language, and Code Llama - Instruct is intended to be safer to use for code assistant and generation applications. **Out-of-Scope Uses** Use in any manner that violates applicable laws or regulations (including trade compliance laws). Use in languages other than English. Use in any other way that is prohibited by the Acceptable Use Policy and Licensing Agreement for Code Llama and its variants. ## Hardware and Software **Training Factors** We used custom training libraries. The training and fine-tuning of the released models have been performed Meta’s Research Super Cluster. **Carbon Footprint** In aggregate, training all 12 Code Llama models required 1400K GPU hours of computation on hardware of type A100-80GB (TDP of 350-400W). Estimated total emissions were 228.55 tCO2eq, 100% of which were offset by Meta’s sustainability program. ## Evaluation Results See evaluations for the main models and detailed ablations in Section 3 and safety evaluations in Section 4 of the research paper. ## Ethical Considerations and Limitations Code Llama and its variants are a new technology that carries risks with use. Testing conducted to date has been in English, and has not covered, nor could it cover all scenarios. For these reasons, as with all LLMs, Code Llama’s potential outputs cannot be predicted in advance, and the model may in some instances produce inaccurate or objectionable responses to user prompts. Therefore, before deploying any applications of Code Llama, developers should perform safety testing and tuning tailored to their specific applications of the model. Please see the Responsible Use Guide available available at [https://ai.meta.com/llama/responsible-use-guide](https://ai.meta.com/llama/responsible-use-guide). <!-- original-model-card end -->
zaenalium/MicroMistral-Indo-400M
zaenalium
2024-01-30T18:10:24Z
10
0
transformers
[ "transformers", "tensorboard", "safetensors", "mistral", "text-generation", "generated_from_trainer", "dataset:MBZUAI/Bactrian-X", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-01-30T15:27:27Z
--- tags: - generated_from_trainer datasets: - MBZUAI/Bactrian-X metrics: - accuracy model-index: - name: Mistral-Indo results: - task: name: Causal Language Modeling type: text-generation dataset: name: MBZUAI/Bactrian-X id type: MBZUAI/Bactrian-X args: id metrics: - name: Accuracy type: accuracy value: 0.2462286333254075 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # Mistral-Indo This model was trained from scratch on the MBZUAI/Bactrian-X id dataset. It achieves the following results on the evaluation set: - Loss: 5.4269 - Accuracy: 0.2462 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 2 - eval_batch_size: 2 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 8 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 20.0 ### Training results ### Framework versions - Transformers 4.37.2 - Pytorch 2.1.2 - Datasets 2.16.1 - Tokenizers 0.15.1
chaitanya-kh/amp_bfloat16_msmarco-distilbert-base-tas-b
chaitanya-kh
2024-01-30T18:04:05Z
47
0
sentence-transformers
[ "sentence-transformers", "safetensors", "distilbert", "feature-extraction", "sentence-similarity", "transformers", "en", "dataset:ms_marco", "license:apache-2.0", "autotrain_compatible", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
sentence-similarity
2024-01-30T06:11:55Z
--- pipeline_tag: sentence-similarity license: apache-2.0 language: "en" tags: - sentence-transformers - feature-extraction - sentence-similarity - transformers datasets: - ms_marco --- # sentence-transformers/msmarco-distilbert-base-tas-b This is a port of the [DistilBert TAS-B Model](https://huggingface.co/sebastian-hofstaetter/distilbert-dot-tas_b-b256-msmarco) to [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and is optimized for the task of semantic search. ## Usage (Sentence-Transformers) Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed: ``` pip install -U sentence-transformers ``` Then you can use the model like this: ```python from sentence_transformers import SentenceTransformer, util query = "How many people live in London?" docs = ["Around 9 Million people live in London", "London is known for its financial district"] #Load the model model = SentenceTransformer('sentence-transformers/msmarco-distilbert-base-tas-b') #Encode query and documents query_emb = model.encode(query) doc_emb = model.encode(docs) #Compute dot score between query and all document embeddings scores = util.dot_score(query_emb, doc_emb)[0].cpu().tolist() #Combine docs & scores doc_score_pairs = list(zip(docs, scores)) #Sort by decreasing score doc_score_pairs = sorted(doc_score_pairs, key=lambda x: x[1], reverse=True) #Output passages & scores for doc, score in doc_score_pairs: print(score, doc) ``` ## Usage (HuggingFace Transformers) Without [sentence-transformers](https://www.SBERT.net), you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings. ```python from transformers import AutoTokenizer, AutoModel import torch #CLS Pooling - Take output from first token def cls_pooling(model_output): return model_output.last_hidden_state[:,0] #Encode text def encode(texts): # Tokenize sentences encoded_input = tokenizer(texts, padding=True, truncation=True, return_tensors='pt') # Compute token embeddings with torch.no_grad(): model_output = model(**encoded_input, return_dict=True) # Perform pooling embeddings = cls_pooling(model_output) return embeddings # Sentences we want sentence embeddings for query = "How many people live in London?" docs = ["Around 9 Million people live in London", "London is known for its financial district"] # Load model from HuggingFace Hub tokenizer = AutoTokenizer.from_pretrained("sentence-transformers/msmarco-distilbert-base-tas-b") model = AutoModel.from_pretrained("sentence-transformers/msmarco-distilbert-base-tas-b") #Encode query and docs query_emb = encode(query) doc_emb = encode(docs) #Compute dot score between query and all document embeddings scores = torch.mm(query_emb, doc_emb.transpose(0, 1))[0].cpu().tolist() #Combine docs & scores doc_score_pairs = list(zip(docs, scores)) #Sort by decreasing score doc_score_pairs = sorted(doc_score_pairs, key=lambda x: x[1], reverse=True) #Output passages & scores for doc, score in doc_score_pairs: print(score, doc) ``` ## Evaluation Results For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name=sentence-transformers/msmarco-distilbert-base-tas-b) ## Full Model Architecture ``` SentenceTransformer( (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: DistilBertModel (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False}) ) ``` ## Citing & Authors Have a look at: [DistilBert TAS-B Model](https://huggingface.co/sebastian-hofstaetter/distilbert-dot-tas_b-b256-msmarco)
AUSTIN2K05/my-pet-ziw
AUSTIN2K05
2024-01-30T17:50:33Z
0
0
null
[ "safetensors", "NxtWave-GenAI-Webinar", "text-to-image", "stable-diffusion", "license:creativeml-openrail-m", "region:us" ]
text-to-image
2024-01-30T17:48:00Z
--- license: creativeml-openrail-m tags: - NxtWave-GenAI-Webinar - text-to-image - stable-diffusion --- ### My-pet-ziw Dreambooth model trained by AUSTIN2K05 following the "Build your own Gen AI model" session by NxtWave. Project Submission Code: GoX19932gAS Sample pictures of this concept: ![0](https://huggingface.co/AUSTIN2K05/my-pet-ziw/resolve/main/sample_images/ziw_mouth_open.png) ![1](https://huggingface.co/AUSTIN2K05/my-pet-ziw/resolve/main/sample_images/ziw_on_big_branch.png)
robertknight/ocrs
robertknight
2024-01-30T17:49:56Z
0
6
null
[ "onnx", "license:cc-by-sa-4.0", "region:us" ]
null
2024-01-30T17:39:21Z
--- license: cc-by-sa-4.0 --- # Ocrs pre-trained models Pre-trained models for the [Ocrs](https://github.com/robertknight/ocrs) OCR engine project. This includes text detection and recognition models available as: - PyTorch checkpoints - ONNX models - [RTen](https://github.com/robertknight/rten) models Training and evaluation scripts and documentation can be found in the [ocrs-models](https://github.com/robertknight/ocrs-models) repository. ## Datasets Models are trained on [HierText](https://github.com/google-research-datasets/hiertext) and synthetic data.
Patcas/codet5Base-Doc-step-1
Patcas
2024-01-30T17:34:15Z
5
0
transformers
[ "transformers", "tensorboard", "safetensors", "t5", "text2text-generation", "generated_from_trainer", "base_model:Salesforce/codet5-base", "base_model:finetune:Salesforce/codet5-base", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text2text-generation
2024-01-30T12:51:14Z
--- license: apache-2.0 base_model: Salesforce/codet5-base tags: - generated_from_trainer model-index: - name: codet5Base-Doc-step-1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # codet5Base-Doc-step-1 This model is a fine-tuned version of [Salesforce/codet5-base](https://huggingface.co/Salesforce/codet5-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.0759 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 1 - eval_batch_size: 1 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | No log | 1.0 | 230 | 1.5475 | | No log | 2.0 | 460 | 1.2669 | | 2.0737 | 3.0 | 690 | 1.1299 | | 2.0737 | 4.0 | 920 | 1.0710 | | 1.0726 | 5.0 | 1150 | 1.0345 | | 1.0726 | 6.0 | 1380 | 1.0217 | | 0.7523 | 7.0 | 1610 | 1.0132 | | 0.7523 | 8.0 | 1840 | 1.0215 | | 0.544 | 9.0 | 2070 | 0.9870 | | 0.544 | 10.0 | 2300 | 0.9926 | | 0.4227 | 11.0 | 2530 | 1.0094 | | 0.4227 | 12.0 | 2760 | 1.0106 | | 0.4227 | 13.0 | 2990 | 1.0087 | | 0.3384 | 14.0 | 3220 | 1.0196 | | 0.3384 | 15.0 | 3450 | 1.0264 | | 0.2754 | 16.0 | 3680 | 1.0283 | | 0.2754 | 17.0 | 3910 | 1.0341 | | 0.2371 | 18.0 | 4140 | 1.0493 | | 0.2371 | 19.0 | 4370 | 1.0506 | | 0.207 | 20.0 | 4600 | 1.0526 | | 0.207 | 21.0 | 4830 | 1.0518 | | 0.1761 | 22.0 | 5060 | 1.0600 | | 0.1761 | 23.0 | 5290 | 1.0644 | | 0.1735 | 24.0 | 5520 | 1.0668 | | 0.1735 | 25.0 | 5750 | 1.0635 | | 0.1735 | 26.0 | 5980 | 1.0726 | | 0.1487 | 27.0 | 6210 | 1.0711 | | 0.1487 | 28.0 | 6440 | 1.0760 | | 0.1451 | 29.0 | 6670 | 1.0757 | | 0.1451 | 30.0 | 6900 | 1.0759 | ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.16.1 - Tokenizers 0.15.1
dhutchings/Reinforce-1
dhutchings
2024-01-30T17:15:55Z
0
0
null
[ "CartPole-v1", "reinforce", "reinforcement-learning", "custom-implementation", "deep-rl-class", "model-index", "region:us" ]
reinforcement-learning
2024-01-30T17:15:47Z
--- tags: - CartPole-v1 - reinforce - reinforcement-learning - custom-implementation - deep-rl-class model-index: - name: Reinforce-1 results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: CartPole-v1 type: CartPole-v1 metrics: - type: mean_reward value: 481.00 +/- 57.00 name: mean_reward verified: false --- # **Reinforce** Agent playing **CartPole-v1** This is a trained model of a **Reinforce** agent playing **CartPole-v1** . To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: https://huggingface.co/deep-rl-course/unit4/introduction
adalib/fate_flow-data-codegen-350M-mono-prefix
adalib
2024-01-30T17:08:05Z
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Salesforce/codegen-350M-mono", "base_model:adapter:Salesforce/codegen-350M-mono", "region:us" ]
null
2024-01-30T17:08:00Z
--- library_name: peft base_model: Salesforce/codegen-350M-mono --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.7.1
chendarren/result
chendarren
2024-01-30T17:03:57Z
0
0
diffusers
[ "diffusers", "tensorboard", "safetensors", "stable-diffusion", "stable-diffusion-diffusers", "text-to-image", "dreambooth", "base_model:runwayml/stable-diffusion-v1-5", "base_model:finetune:runwayml/stable-diffusion-v1-5", "license:creativeml-openrail-m", "autotrain_compatible", "endpoints_compatible", "diffusers:StableDiffusionPipeline", "region:us" ]
text-to-image
2024-01-30T13:35:21Z
--- license: creativeml-openrail-m base_model: runwayml/stable-diffusion-v1-5 instance_prompt: a photo of sks tank tags: - stable-diffusion - stable-diffusion-diffusers - text-to-image - diffusers - dreambooth inference: true --- # DreamBooth - chendarren/result This is a dreambooth model derived from runwayml/stable-diffusion-v1-5. The weights were trained on a photo of sks tank using [DreamBooth](https://dreambooth.github.io/). You can find some example images in the following. DreamBooth for the text encoder was enabled: False.
adalib/evaluate-data-CodeGPT-small-py-prefix
adalib
2024-01-30T16:54:31Z
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:microsoft/CodeGPT-small-py", "base_model:adapter:microsoft/CodeGPT-small-py", "region:us" ]
null
2024-01-30T16:54:27Z
--- library_name: peft base_model: microsoft/CodeGPT-small-py --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.7.1
helenai/mistralai-Mistral-7B-v0.1-ov
helenai
2024-01-30T16:54:06Z
2
0
transformers
[ "transformers", "openvino", "mistral", "text-generation", "en", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
2024-01-30T14:42:54Z
--- language: - en tags: - openvino --- # mistralai/Mistral-7B-v0.1 This is the [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) model converted to [OpenVINO](https://openvino.ai), for accellerated inference. An example of how to do inference on this model: ```python from optimum.intel.openvino import OVModelForCausalLM from transformers import AutoTokenizer, pipeline # model_id should be set to either a local directory or a model available on the HuggingFace hub. model_id = "helenai/mistralai-Mistral-7B-v0.1-ov" tokenizer = AutoTokenizer.from_pretrained(model_id) model = OVModelForCausalLM.from_pretrained(model_id) pipe = pipeline("text-generation", model=model, tokenizer=tokenizer) result = pipe("My name is Julien and I like to") print(result) ```
LoneStriker/CodeLlama-70b-hf-2.65bpw-h6-exl2
LoneStriker
2024-01-30T16:51:49Z
5
0
transformers
[ "transformers", "pytorch", "safetensors", "llama", "text-generation", "llama-2", "code", "arxiv:2308.12950", "license:llama2", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-01-30T09:55:29Z
--- language: - code pipeline_tag: text-generation tags: - llama-2 license: llama2 --- # **Code Llama** Code Llama is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the repository for the base 70B version in the Hugging Face Transformers format. This model is designed for general code synthesis and understanding. Links to other models can be found in the index at the bottom. | | Base Model | Python | Instruct | | --- | ----------------------------------------------------------------------------- | ------------------------------------------------------------------------------------------- | ----------------------------------------------------------------------------------------------- | | 7B | [codellama/CodeLlama-7b-hf](https://huggingface.co/codellama/CodeLlama-7b-hf) | [codellama/CodeLlama-7b-Python-hf](https://huggingface.co/codellama/CodeLlama-7b-Python-hf) | [codellama/CodeLlama-7b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-7b-Instruct-hf) | | 13B | [codellama/CodeLlama-13b-hf](https://huggingface.co/codellama/CodeLlama-13b-hf) | [codellama/CodeLlama-13b-Python-hf](https://huggingface.co/codellama/CodeLlama-13b-Python-hf) | [codellama/CodeLlama-13b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-13b-Instruct-hf) | | 34B | [codellama/CodeLlama-34b-hf](https://huggingface.co/codellama/CodeLlama-34b-hf) | [codellama/CodeLlama-34b-Python-hf](https://huggingface.co/codellama/CodeLlama-34b-Python-hf) | [codellama/CodeLlama-34b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-34b-Instruct-hf) | | 70B | [codellama/CodeLlama-70b-hf](https://huggingface.co/codellama/CodeLlama-70b-hf) | [codellama/CodeLlama-70b-Python-hf](https://huggingface.co/codellama/CodeLlama-70b-Python-hf) | [codellama/CodeLlama-70b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-70b-Instruct-hf) | ## Model Use To use this model, please make sure to install `transformers`. ```bash pip install transformers accelerate ``` Model capabilities: - [x] Code completion. - [ ] Infilling. - [ ] Instructions / chat. - [ ] Python specialist. ## Model Details *Note: Use of this model is governed by the Meta license. Meta developed and publicly released the Code Llama family of large language models (LLMs). **Model Developers** Meta **Variations** Code Llama comes in four model sizes, and three variants: * Code Llama: base models designed for general code synthesis and understanding * Code Llama - Python: designed specifically for Python * Code Llama - Instruct: for instruction following and safer deployment All variants are available in sizes of 7B, 13B, 34B, and 70B parameters. **This repository contains the base version of the 70B parameters model.** **Input** Models input text only. **Output** Models generate text only. **Model Architecture** Code Llama is an auto-regressive language model that uses an optimized transformer architecture. It was fine-tuned with up to 16k tokens and supports up to 100k tokens at inference time. **Model Dates** Code Llama and its variants have been trained between January 2023 and January 2024. **Status** This is a static model trained on an offline dataset. Future versions of Code Llama - Instruct will be released as we improve model safety with community feedback. **License** A custom commercial license is available at: [https://ai.meta.com/resources/models-and-libraries/llama-downloads/](https://ai.meta.com/resources/models-and-libraries/llama-downloads/) **Research Paper** More information can be found in the paper "[Code Llama: Open Foundation Models for Code](https://ai.meta.com/research/publications/code-llama-open-foundation-models-for-code/)" or its [arXiv page](https://arxiv.org/abs/2308.12950). ## Intended Use **Intended Use Cases** Code Llama and its variants are intended for commercial and research use in English and relevant programming languages. The base model Code Llama can be adapted for a variety of code synthesis and understanding tasks, Code Llama - Python is designed specifically to handle the Python programming language, and Code Llama - Instruct is intended to be safer to use for code assistant and generation applications. **Out-of-Scope Uses** Use in any manner that violates applicable laws or regulations (including trade compliance laws). Use in languages other than English. Use in any other way that is prohibited by the Acceptable Use Policy and Licensing Agreement for Code Llama and its variants. ## Hardware and Software **Training Factors** We used custom training libraries. The training and fine-tuning of the released models have been performed Meta’s Research Super Cluster. **Carbon Footprint** In aggregate, training all 12 Code Llama models required 1400K GPU hours of computation on hardware of type A100-80GB (TDP of 350-400W). Estimated total emissions were 228.55 tCO2eq, 100% of which were offset by Meta’s sustainability program. ## Evaluation Results See evaluations for the main models and detailed ablations in Section 3 and safety evaluations in Section 4 of the research paper. ## Ethical Considerations and Limitations Code Llama and its variants are a new technology that carries risks with use. Testing conducted to date has been in English, and has not covered, nor could it cover all scenarios. For these reasons, as with all LLMs, Code Llama’s potential outputs cannot be predicted in advance, and the model may in some instances produce inaccurate or objectionable responses to user prompts. Therefore, before deploying any applications of Code Llama, developers should perform safety testing and tuning tailored to their specific applications of the model. Please see the Responsible Use Guide available available at [https://ai.meta.com/llama/responsible-use-guide](https://ai.meta.com/llama/responsible-use-guide).
jondurbin/bagel-dpo-7b-v0.1
jondurbin
2024-01-30T16:49:41Z
1,416
42
transformers
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "dataset:ai2_arc", "dataset:unalignment/spicy-3.1", "dataset:codeparrot/apps", "dataset:facebook/belebele", "dataset:boolq", "dataset:jondurbin/cinematika-v0.1", "dataset:drop", "dataset:lmsys/lmsys-chat-1m", "dataset:TIGER-Lab/MathInstruct", "dataset:cais/mmlu", "dataset:Muennighoff/natural-instructions", "dataset:openbookqa", "dataset:piqa", "dataset:Vezora/Tested-22k-Python-Alpaca", "dataset:cakiki/rosetta-code", "dataset:Open-Orca/SlimOrca", "dataset:spider", "dataset:squad_v2", "dataset:migtissera/Synthia-v1.3", "dataset:datasets/winogrande", "dataset:nvidia/HelpSteer", "dataset:Intel/orca_dpo_pairs", "dataset:unalignment/toxic-dpo-v0.1", "dataset:jondurbin/truthy-dpo-v0.1", "dataset:allenai/ultrafeedback_binarized_cleaned", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2023-12-13T12:12:22Z
--- license: apache-2.0 datasets: - ai2_arc - unalignment/spicy-3.1 - codeparrot/apps - facebook/belebele - boolq - jondurbin/cinematika-v0.1 - drop - lmsys/lmsys-chat-1m - TIGER-Lab/MathInstruct - cais/mmlu - Muennighoff/natural-instructions - openbookqa - piqa - Vezora/Tested-22k-Python-Alpaca - cakiki/rosetta-code - Open-Orca/SlimOrca - spider - squad_v2 - migtissera/Synthia-v1.3 - datasets/winogrande - nvidia/HelpSteer - Intel/orca_dpo_pairs - unalignment/toxic-dpo-v0.1 - jondurbin/truthy-dpo-v0.1 - allenai/ultrafeedback_binarized_cleaned --- # A bagel, with everything ![bagel](bagel.png) ## Overview This is the DPO'd version of https://huggingface.co/jondurbin/bagel-7b-v0.1 If you are getting too many AALLM or other refusals, even with explicitly human system prompts, you may want to try the non-DPO version. ## Benchmarks I ran these against the latest main branch of lm-evaluation-harness (and opencompass/FastChat for agieval and mt-bench), since batch size/etc effects score for some benchmarks. | model | arc_challenge | boolq | gsm8k | hellaswag | mmlu | openbookqa | piqa | truthful_qa | winogrande | | --- | --- | --- | --- | --- | --- | --- | --- | --- | --- | | bagel | __0.6715__ | 0.8813 | __0.5618__ | 0.8397 | __0.6408__ | __0.51__ | __0.8406__ | __0.6275__ | __0.7561__ | | openhermes-2.5 | 0.6476 | __0.8835__ | 0.4852 | __0.8414__ | 0.6347 | 0.498 | 0.8400 | 0.5295 | 0.7443 | MT-Bench: ``` ########## First turn ########## score model turn bagel-7b-v0.1 1 7.60625 ########## Second turn ########## score model turn bagel-7b-v0.1 2 7.00625 ########## Average ########## score model bagel-7b-v0.1 7.30625 ``` ## Data selection. The first step in the process is creating a dataset. In this case, we're actually creating a composite dataset, consisting of both supervised fine-tuning data (SFT) and direct preference optimization (DPO) data. All instruction data, that is, data that is not plain text (like project Gutenberg and items from Cinematika) or DPO, is converted into ShareGPT format so it's easier to work with. See the corresponding code in `bagel/data_sources/*.py` for full implementation for each data source. Deduplication is done by creating a uuid v5 of the instruction/text, then only adding items not previously seen (where datasets are loaded in order of the confidence score I assign them). This means that if an instruction is in data source "Foo" with confidence 4 as well as in data source "Bar" with confidence score 2, only the entry from "Foo" will be taken. ### SFT data sources *Yes, you will see benchmark names in the list, but this only uses the train splits, and a decontamination by cosine similarity is performed at the end as a sanity check* - [ai2_arc](https://huggingface.co/datasets/ai2_arc) - Abstraction and reasoning dataset, useful in measuring "intelligence" to a certain extent. - [airoboros](https://huggingface.co/datasets/unalignment/spicy-3.1) - Variety of categories of synthetic instructions generated by gpt-4. - [apps](https://huggingface.co/datasets/codeparrot/apps) - Python coding dataset with 10k problems. - [belebele](https://huggingface.co/datasets/facebook/belebele) - Multi-lingual reading comprehension dataset. - [boolq](https://huggingface.co/datasets/boolq) - Corpus of yes/no questions (which can be surprisingly difficult for AI to answer apparently?) - [cinematika](https://huggingface.co/datasets/jondurbin/cinematika-v0.1) (instruction and plain text) - RP-style data synthesized from movie scripts so the model isn't quite as boring as it otherwise would be. - [drop](https://huggingface.co/datasets/drop) - More reading comprehension. - [gutenberg](https://www.gutenberg.org/) (plain text) - Books/plain text, again to make the model less boring, only a handful of examples supported by [chapterize](https://github.com/JonathanReeve/chapterize) - [lmsys_chat_1m](https://huggingface.co/datasets/lmsys/lmsys-chat-1m) (only gpt-4 items, also used for DPO) - Chats collected by the lmsys chat arena, containing a wide variety of chats with various models. - [mathinstruct](https://huggingface.co/datasets/TIGER-Lab/MathInstruct) - Composite dataset with a variety of math-related tasks and problem/question formats. - [mmlu](https://huggingface.co/datasets/cais/mmlu) - Massive Multitask Language Understanding - a wide variety of questions about various subject matters. - [natural_instructions](https://huggingface.co/datasets/Muennighoff/natural-instructions) - Millions of instructions from 1600+ task categories (sampled down substantially, stratified by task type) - [openbookqa](https://huggingface.co/datasets/openbookqa) - Question answering dataset. - [piqa](https://huggingface.co/datasets/piqa) - Phyiscal interaction question answering. - [python_alpaca](https://huggingface.co/datasets/Vezora/Tested-22k-Python-Alpaca) - Python instruction response pairs, validated as functional. - [rosetta_code](https://huggingface.co/datasets/cakiki/rosetta-code) - Code problems and solutions in a variety of programming languages taken from rosettacode.org. - [slimorca](https://huggingface.co/datasets/Open-Orca/SlimOrca) - Collection of ~500k gpt-4 verified chats from OpenOrca. - [spider](https://huggingface.co/datasets/spider) - SQL-targeted dataset. - [squad_v2](https://huggingface.co/datasets/squad_v2) - Contextual question answering (RAG). - [synthia](https://huggingface.co/datasets/migtissera/Synthia-v1.3) - GPT-4 generated data using advanced prompting from Migel Tissera. - [winogrande](https://huggingface.co/datasets/winogrande) - Fill in the blank style prompts. ### DPO data sources - [airoboros 3.1](https://huggingface.co/datasets/unalignment/spicy-3.1) vs [airoboros 2.2.1](https://huggingface.co/datasets/jondurbin/airoboros-gpt4-1.4.1) - The creative/writing tasks from airoboros-2.2.1 were re-generated using gpt4-0314 and a custom prompt to get longer, more creative, less clichè responses for airoboros 3.1, so we can use the shorter/boring version as the "rejected" value and the rerolled response as "chosen" - [helpsteer](https://huggingface.co/datasets/nvidia/HelpSteer) - Really neat dataset provided by the folks at NVidia with human annotation across a variety of metrics. Only items with the highest "correctness" value were used for DPO here, with the highest scoring output as "chosen" and random lower scoring value as "rejected" - [orca_dpo_pairs](https://huggingface.co/datasets/Intel/orca_dpo_pairs) - Another interesting dataset by Intel, which provides various DPO pairs generated from prompts included in the SlimOrca dataset. - [toxic-dpo](https://huggingface.co/datasets/unalignment/toxic-dpo-v0.1) - __*highly toxic and potentially illegal content!*__ De-censorship, for academic and lawful purposes only, of course. Generated by llama-2-70b via prompt engineering. - [truthy](https://huggingface.co/datasets/jondurbin/truthy-dpo-v0.1) - DPO pairs meant to increase truthfulness of the model, e.g. common misconceptions, differentiate between AI assistants and roleplayed human in terms of corporeal awareness/locality/etc. - [ultrafeedback](https://huggingface.co/datasets/allenai/ultrafeedback_binarized_cleaned) - One of the bits of magic behind the Zephyr model. Only the items with a chosen score of 8 or higher were included. Only the train splits were used (if a split was provided), and an additional pass of decontamination is performed using approximate nearest neighbor search (via faiss). ### Total dataset size The deduplicated and decontamined list of instructions contains 1,671,822 items: - 1,602,217 SFT/instructions - 59,247 DPO pairs - 1606 with both SFT and DPO data Keep in mind, this number becomes 4x larger when applying the various prompt formats. ## Prompt formatting In sticking with the theme of the bagel, I didn't want to use a single prompt format, so I used 4 - vicuna, llama-2, alpaca, and chat-ml (sorta). I also didn't want to randomly select a single prompt format for each item (hoping each instruction would generalize more when used in a variety of prompt formats), so each instruction is actually converted into every prompt format. This means each epoch of our fine-tune is really basically 4 epochs. So, for the fine-tunes, I would recommend only doing 1 epoch (or 0.75 epochs). I am testing with a single epoch using a relatively low learning rate. ### Alpaca (sort of) ``` Below is an instruction that describes a task. Write a response that appropriately completes the request. ### Instruction: {system prompt, if provided} {instruction} ### Response: ``` The main difference here is that because of the dataset formatting and variety of data sources, it would have been much to tedious to add an `### Input:` block, so the inputs are just in the instruction section. ### Vicuna ``` {system prompt, if provided, randomly defaulting to "A chat between a user and an unbiased, uncensored assistant."} USER: {instruction} ASSISTANT: ``` ### ChatML (sort of) I don't really understand the point of having special tokens for `<|im_start|>` and `<|im_end|>`, because in practice they just act as BOS and EOS tokens (but, please correct me if I'm wrong). So, instead of: ```text {bos}<|im_start|>{role} {text} <|im_end|>{eos} ``` I just changed it to: ```text {bos}{role} {text} {eos} ``` In practice, this would mean tokenization code like such: ```python tokenizer = AutoTokenizer.from_pretrained('mistralai/mistral-7b-v0.1') input_str = f"""system You are a goat. {tokenizer.eos_token} {tokenizer.bos_token}user Tell me how to fry an egg. {tokenizer.eos_token} {tokenizer.bos_token}assistant """ inputs = tokenizer(input_str, return_tensors="pt") ``` If you *really* want to use `<|im_start|>` and `<|im_end|>`, just update your `tokenizer_config.json` to use `<|im_start|>` instead of `<s>` and `<|im_end|>` instead of `</s>` and when tokenizing. And if you still don't like what I've done to this chat-ml-ish format, feel free to cry into your pillow or fork the code and do a new fine-tune. ### Llama-2 chat ``` [INST] <<SYS>> {system} <</SYS>> {instruction} [/INST] ``` ## Fine tuning ### SFT phase An example for mistral-7b: *Note: I actually used my fork of [qlora](https://github.com/jondurbin/qlora)'s `train.py` for this, but I'm porting it to a minified version here, not tested yet!* *More notes: I stopped the SFT phase around 50% because of budget constraints.* ```bash export BASE_DIR=/workspace export WANDB_API_KEY=[redacted] export WANDB_PROJECT=bagel-7b-v0.1 # Run the pretraining. accelerate launch bagel/tune/sft.py \ --model_name_or_path $BASE_DIR/mistral-7b \ --final_output_dir $BASE_DIR/$WANDB_PROJECT \ --output_dir $BASE_DIR/$WANDB_PROJECT-workdir \ --num_train_epochs 1 \ --logging_steps 1 \ --save_strategy steps \ --save_steps 200 \ --save_total_limit 5 \ --data_seed 42 \ --evaluation_strategy steps \ --eval_dataset_size 0.0006 \ --eval_steps 200 \ --max_new_tokens 4096 \ --dataloader_num_workers 3 \ --logging_strategy steps \ --remove_unused_columns False \ --do_train \ --full_finetune \ --bf16 \ --bits 16 \ --optim adamw_torch \ --lr_scheduler_type linear \ --dataset $BASE_DIR/bagel/bagel-input-output-v0.1.parquet \ --dataset_format input-output \ --model_max_len 4096 \ --per_device_train_batch_size 8 \ --learning_rate 3.5e-7 \ --warmup_ratio 0.005 \ --adam_beta2 0.999 \ --max_grad_norm 0.3 \ --weight_decay 0.001 \ --seed 42 \ --report_to wandb \ --gradient_checkpointing True \ --gradient_accumulation_steps 4 \ --skip_excess_length False \ --ddp_find_unused_parameters False \ --use_flash_attention_2 \ --deepspeed deepspeed.json ``` Deepspeed configuration: ```json { "gradient_accumulation_steps": "auto", "gradient_clipping": "auto", "train_batch_size": "auto", "train_micro_batch_size_per_gpu": "auto", "bf16": { "enabled": true }, "zero_optimization": { "stage": 2, "contiguous_gradients": true, "overlap_comm": true, "reduce_scatter": true, "reduce_bucket_size": 5e8, "allgather_bucket_size": 5e8 } } ``` ### DPO phase An example of the DPO phase for mistral-7b (requires first running the SFT): ```bash export BASE_DIR=/mnt/data export WANDB_API_KEY=[redacted] export WANDB_PROJECT=bagel-dpo-7b-v0.1 accelerate launch bagel/tune/dpo.py \ --model_name_or_path bagel-7b-v0.1 \ --learning_rate 3e-7 \ --per_device_train_batch_size 2 \ --gradient_accumulation_steps 4 \ --max_length 4096 \ --max_prompt_length 1024 \ --max_target_length 3092 \ --num_train_epochs 3 \ --report_to wandb \ --gradient_checkpointing true \ --use_flash_attention_2 true \ --dataset $BASE_DIR/bagel/bagel-dpo-v0.1.parquet \ --eval_steps 5 \ --eval_dataset_size 0.03 \ --workdir $BASE_DIR/$WANDB_PROJECT-workdir \ --output_dir $BASE_DIR/$WANDB_PROJECT \ --deepspeed deepspeed.json \ --save_steps 25 \ --save_total_limit 5 ```
phihung/ppo-LunarLander-v2
phihung
2024-01-30T16:49:28Z
1
0
stable-baselines3
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2024-01-30T16:10:23Z
--- library_name: stable-baselines3 tags: - LunarLander-v2 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 metrics: - type: mean_reward value: 285.81 +/- 14.08 name: mean_reward verified: false --- # **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
LoneStriker/CodeLlama-70b-hf-4.65bpw-h6-exl2
LoneStriker
2024-01-30T16:41:00Z
4
0
transformers
[ "transformers", "pytorch", "safetensors", "llama", "text-generation", "llama-2", "code", "arxiv:2308.12950", "license:llama2", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-01-30T10:34:03Z
--- language: - code pipeline_tag: text-generation tags: - llama-2 license: llama2 --- # **Code Llama** Code Llama is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the repository for the base 70B version in the Hugging Face Transformers format. This model is designed for general code synthesis and understanding. Links to other models can be found in the index at the bottom. | | Base Model | Python | Instruct | | --- | ----------------------------------------------------------------------------- | ------------------------------------------------------------------------------------------- | ----------------------------------------------------------------------------------------------- | | 7B | [codellama/CodeLlama-7b-hf](https://huggingface.co/codellama/CodeLlama-7b-hf) | [codellama/CodeLlama-7b-Python-hf](https://huggingface.co/codellama/CodeLlama-7b-Python-hf) | [codellama/CodeLlama-7b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-7b-Instruct-hf) | | 13B | [codellama/CodeLlama-13b-hf](https://huggingface.co/codellama/CodeLlama-13b-hf) | [codellama/CodeLlama-13b-Python-hf](https://huggingface.co/codellama/CodeLlama-13b-Python-hf) | [codellama/CodeLlama-13b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-13b-Instruct-hf) | | 34B | [codellama/CodeLlama-34b-hf](https://huggingface.co/codellama/CodeLlama-34b-hf) | [codellama/CodeLlama-34b-Python-hf](https://huggingface.co/codellama/CodeLlama-34b-Python-hf) | [codellama/CodeLlama-34b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-34b-Instruct-hf) | | 70B | [codellama/CodeLlama-70b-hf](https://huggingface.co/codellama/CodeLlama-70b-hf) | [codellama/CodeLlama-70b-Python-hf](https://huggingface.co/codellama/CodeLlama-70b-Python-hf) | [codellama/CodeLlama-70b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-70b-Instruct-hf) | ## Model Use To use this model, please make sure to install `transformers`. ```bash pip install transformers accelerate ``` Model capabilities: - [x] Code completion. - [ ] Infilling. - [ ] Instructions / chat. - [ ] Python specialist. ## Model Details *Note: Use of this model is governed by the Meta license. Meta developed and publicly released the Code Llama family of large language models (LLMs). **Model Developers** Meta **Variations** Code Llama comes in four model sizes, and three variants: * Code Llama: base models designed for general code synthesis and understanding * Code Llama - Python: designed specifically for Python * Code Llama - Instruct: for instruction following and safer deployment All variants are available in sizes of 7B, 13B, 34B, and 70B parameters. **This repository contains the base version of the 70B parameters model.** **Input** Models input text only. **Output** Models generate text only. **Model Architecture** Code Llama is an auto-regressive language model that uses an optimized transformer architecture. It was fine-tuned with up to 16k tokens and supports up to 100k tokens at inference time. **Model Dates** Code Llama and its variants have been trained between January 2023 and January 2024. **Status** This is a static model trained on an offline dataset. Future versions of Code Llama - Instruct will be released as we improve model safety with community feedback. **License** A custom commercial license is available at: [https://ai.meta.com/resources/models-and-libraries/llama-downloads/](https://ai.meta.com/resources/models-and-libraries/llama-downloads/) **Research Paper** More information can be found in the paper "[Code Llama: Open Foundation Models for Code](https://ai.meta.com/research/publications/code-llama-open-foundation-models-for-code/)" or its [arXiv page](https://arxiv.org/abs/2308.12950). ## Intended Use **Intended Use Cases** Code Llama and its variants are intended for commercial and research use in English and relevant programming languages. The base model Code Llama can be adapted for a variety of code synthesis and understanding tasks, Code Llama - Python is designed specifically to handle the Python programming language, and Code Llama - Instruct is intended to be safer to use for code assistant and generation applications. **Out-of-Scope Uses** Use in any manner that violates applicable laws or regulations (including trade compliance laws). Use in languages other than English. Use in any other way that is prohibited by the Acceptable Use Policy and Licensing Agreement for Code Llama and its variants. ## Hardware and Software **Training Factors** We used custom training libraries. The training and fine-tuning of the released models have been performed Meta’s Research Super Cluster. **Carbon Footprint** In aggregate, training all 12 Code Llama models required 1400K GPU hours of computation on hardware of type A100-80GB (TDP of 350-400W). Estimated total emissions were 228.55 tCO2eq, 100% of which were offset by Meta’s sustainability program. ## Evaluation Results See evaluations for the main models and detailed ablations in Section 3 and safety evaluations in Section 4 of the research paper. ## Ethical Considerations and Limitations Code Llama and its variants are a new technology that carries risks with use. Testing conducted to date has been in English, and has not covered, nor could it cover all scenarios. For these reasons, as with all LLMs, Code Llama’s potential outputs cannot be predicted in advance, and the model may in some instances produce inaccurate or objectionable responses to user prompts. Therefore, before deploying any applications of Code Llama, developers should perform safety testing and tuning tailored to their specific applications of the model. Please see the Responsible Use Guide available available at [https://ai.meta.com/llama/responsible-use-guide](https://ai.meta.com/llama/responsible-use-guide).
potsawee/deberta-v3-large-mnli
potsawee
2024-01-30T16:37:55Z
29,556
8
transformers
[ "transformers", "pytorch", "deberta-v2", "text-classification", "en", "dataset:multi_nli", "arxiv:2303.08896", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2023-07-18T15:21:49Z
--- license: apache-2.0 datasets: - multi_nli language: - en pipeline_tag: text-classification --- # DeBERTa-v3 (large) fine-tuned to Multi-NLI (MNLI) This model is for Textual Entailment (aka NLI), i.e., predict whether `textA` is supported by `textB`. More specifically, it's a 2-way classification where the relationship between `textA` and `textB` can be **entail, neutral, contradict**. - Input: (`textA`, `textB`) - Output: prob(entail), prob(contradict) Note that during training, all 3 labels (entail, neural, contradict) were used. But for this model, the neural output head has been removed. ## Model Details - Base model: [deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) - Training data: [MNLI](https://huggingface.co/datasets/multi_nli) - Training details: num_epochs = 3, batch_size = 16, `textA=hypothesis`, `textB=premise` ## Example ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification tokenizer = AutoTokenizer.from_pretrained("potsawee/deberta-v3-large-mnli") model = AutoModelForSequenceClassification.from_pretrained("potsawee/deberta-v3-large-mnli") textA = "Kyle Walker has a personal issue" textB = "Kyle Walker will remain Manchester City captain following reports about his private life, says boss Pep Guardiola." inputs = tokenizer.batch_encode_plus( batch_text_or_text_pairs=[(textA, textB)], add_special_tokens=True, return_tensors="pt", ) logits = model(**inputs).logits # neutral is already removed probs = torch.softmax(logits, dim=-1)[0] # probs = [0.7080, 0.2920], meaning that prob(entail) = 0.708, prob(contradict) = 0.292 ``` ## Citation ```bibtex @article{manakul2023selfcheckgpt, title={Selfcheckgpt: Zero-resource black-box hallucination detection for generative large language models}, author={Manakul, Potsawee and Liusie, Adian and Gales, Mark JF}, journal={arXiv preprint arXiv:2303.08896}, year={2023} } ```
JorgePVNV/mi_modelo_test_1
JorgePVNV
2024-01-30T16:35:47Z
14
0
transformers
[ "transformers", "tensorboard", "safetensors", "roberta", "text-classification", "generated_from_trainer", "base_model:pysentimiento/roberta-es-sentiment", "base_model:finetune:pysentimiento/roberta-es-sentiment", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-01-30T12:08:18Z
--- base_model: pysentimiento/roberta-es-sentiment tags: - generated_from_trainer metrics: - f1 model-index: - name: mi_modelo_test_1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # mi_modelo_test_1 This model is a fine-tuned version of [pysentimiento/roberta-es-sentiment](https://huggingface.co/pysentimiento/roberta-es-sentiment) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.6274 - F1: 0.7635 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | F1 | |:-------------:|:-----:|:----:|:---------------:|:------:| | 0.6231 | 1.0 | 26 | 0.6274 | 0.7635 | | 0.1 | 2.0 | 52 | 0.8851 | 0.7358 | | 0.0076 | 3.0 | 78 | 1.0922 | 0.7221 | | 0.0041 | 4.0 | 104 | 1.1844 | 0.7556 | | 0.0034 | 5.0 | 130 | 1.1963 | 0.7505 | ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.16.1 - Tokenizers 0.15.1
TTNVXX/CartoonOrNot
TTNVXX
2024-01-30T16:34:28Z
197
0
transformers
[ "transformers", "safetensors", "swin", "image-classification", "autotrain", "dataset:autotrain-oaqu7-o86bm/autotrain-data", "autotrain_compatible", "endpoints_compatible", "region:us" ]
image-classification
2024-01-30T16:33:56Z
--- tags: - autotrain - image-classification widget: - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/tiger.jpg example_title: Tiger - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/teapot.jpg example_title: Teapot - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/palace.jpg example_title: Palace datasets: - autotrain-oaqu7-o86bm/autotrain-data --- # Model Trained Using AutoTrain - Problem type: Image Classification ## Validation Metricsg loss: 0.039548490196466446 f1: 0.9896907216494846 precision: 1.0 recall: 0.9795918367346939 auc: 0.9991376832423111 accuracy: 0.9916666666666667
mergisi/falcon7binstruct_text_to_sql_optimized_v9
mergisi
2024-01-30T16:32:42Z
2
0
peft
[ "peft", "tensorboard", "safetensors", "trl", "sft", "generated_from_trainer", "base_model:vilsonrodrigues/falcon-7b-instruct-sharded", "base_model:adapter:vilsonrodrigues/falcon-7b-instruct-sharded", "license:apache-2.0", "region:us" ]
null
2024-01-30T12:48:42Z
--- license: apache-2.0 library_name: peft tags: - trl - sft - generated_from_trainer base_model: vilsonrodrigues/falcon-7b-instruct-sharded model-index: - name: falcon7binstruct_text_to_sql_optimized_v9 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # falcon7binstruct_text_to_sql_optimized_v9 This model is a fine-tuned version of [vilsonrodrigues/falcon-7b-instruct-sharded](https://huggingface.co/vilsonrodrigues/falcon-7b-instruct-sharded) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 32 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 8 - total_train_batch_size: 256 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: constant - lr_scheduler_warmup_ratio: 0.03 - training_steps: 180 - mixed_precision_training: Native AMP ### Training results ### Framework versions - PEFT 0.8.2.dev0 - Transformers 4.37.2 - Pytorch 2.1.0+cu121 - Datasets 2.16.1 - Tokenizers 0.15.1
mmnga/OrionStarAI-Orion-14B-LongChat-gguf
mmnga
2024-01-30T16:23:45Z
148
1
null
[ "gguf", "license:other", "endpoints_compatible", "region:us" ]
null
2024-01-29T18:03:07Z
--- license: other license_name: orion-14b-series-models-community-license license_link: LICENSE --- # OrionStarAI-Orion-14B-LongChat-gguf [OrionStarAIさんが公開しているOrion-14B-LongChat](https://huggingface.co/OrionStarAI/Orion-14B-LongChat)のggufフォーマット変換版です。 ## Licence 元モデルのライセンス項目をご確認ください。 [orion-14b-series-models-community-license](https://huggingface.co/mmnga/OrionStarAI-Orion-14B-LongChat-gguf/blob/main/LICENSE) 他のモデル [mmnga/OrionStarAI-Orion-14B-LongChat-gguf](https://huggingface.co/mmnga/OrionStarAI-Orion-14B-LongChat-gguf) [mmnga/OrionStarAI-Orion-14B-Chat-RAG-gguf](https://huggingface.co/mmnga/OrionStarAI-Orion-14B-Chat-RAG-gguf) ## Usage ``` git clone https://github.com/ggerganov/llama.cpp.git cd llama.cpp make -j ./main -m 'OrionStarAI-Orion-14B-LongChat-q4_0.gguf' -p "User: 今晩の夕食のレシピを考えて \n Orion-14B: " -n 128 --frequency-penalty 0.5 --frequency-penalty 0.5 --top-k 5 --top-p 0.9 -e ```
LoneStriker/CodeLlama-70b-hf-2.4bpw-h6-exl2
LoneStriker
2024-01-30T16:23:14Z
6
1
transformers
[ "transformers", "pytorch", "safetensors", "llama", "text-generation", "llama-2", "code", "arxiv:2308.12950", "license:llama2", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-01-30T16:12:24Z
--- language: - code pipeline_tag: text-generation tags: - llama-2 license: llama2 --- # **Code Llama** Code Llama is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the repository for the base 70B version in the Hugging Face Transformers format. This model is designed for general code synthesis and understanding. Links to other models can be found in the index at the bottom. | | Base Model | Python | Instruct | | --- | ----------------------------------------------------------------------------- | ------------------------------------------------------------------------------------------- | ----------------------------------------------------------------------------------------------- | | 7B | [codellama/CodeLlama-7b-hf](https://huggingface.co/codellama/CodeLlama-7b-hf) | [codellama/CodeLlama-7b-Python-hf](https://huggingface.co/codellama/CodeLlama-7b-Python-hf) | [codellama/CodeLlama-7b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-7b-Instruct-hf) | | 13B | [codellama/CodeLlama-13b-hf](https://huggingface.co/codellama/CodeLlama-13b-hf) | [codellama/CodeLlama-13b-Python-hf](https://huggingface.co/codellama/CodeLlama-13b-Python-hf) | [codellama/CodeLlama-13b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-13b-Instruct-hf) | | 34B | [codellama/CodeLlama-34b-hf](https://huggingface.co/codellama/CodeLlama-34b-hf) | [codellama/CodeLlama-34b-Python-hf](https://huggingface.co/codellama/CodeLlama-34b-Python-hf) | [codellama/CodeLlama-34b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-34b-Instruct-hf) | | 70B | [codellama/CodeLlama-70b-hf](https://huggingface.co/codellama/CodeLlama-70b-hf) | [codellama/CodeLlama-70b-Python-hf](https://huggingface.co/codellama/CodeLlama-70b-Python-hf) | [codellama/CodeLlama-70b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-70b-Instruct-hf) | ## Model Use To use this model, please make sure to install `transformers`. ```bash pip install transformers accelerate ``` Model capabilities: - [x] Code completion. - [ ] Infilling. - [ ] Instructions / chat. - [ ] Python specialist. ## Model Details *Note: Use of this model is governed by the Meta license. Meta developed and publicly released the Code Llama family of large language models (LLMs). **Model Developers** Meta **Variations** Code Llama comes in four model sizes, and three variants: * Code Llama: base models designed for general code synthesis and understanding * Code Llama - Python: designed specifically for Python * Code Llama - Instruct: for instruction following and safer deployment All variants are available in sizes of 7B, 13B, 34B, and 70B parameters. **This repository contains the base version of the 70B parameters model.** **Input** Models input text only. **Output** Models generate text only. **Model Architecture** Code Llama is an auto-regressive language model that uses an optimized transformer architecture. It was fine-tuned with up to 16k tokens and supports up to 100k tokens at inference time. **Model Dates** Code Llama and its variants have been trained between January 2023 and January 2024. **Status** This is a static model trained on an offline dataset. Future versions of Code Llama - Instruct will be released as we improve model safety with community feedback. **License** A custom commercial license is available at: [https://ai.meta.com/resources/models-and-libraries/llama-downloads/](https://ai.meta.com/resources/models-and-libraries/llama-downloads/) **Research Paper** More information can be found in the paper "[Code Llama: Open Foundation Models for Code](https://ai.meta.com/research/publications/code-llama-open-foundation-models-for-code/)" or its [arXiv page](https://arxiv.org/abs/2308.12950). ## Intended Use **Intended Use Cases** Code Llama and its variants are intended for commercial and research use in English and relevant programming languages. The base model Code Llama can be adapted for a variety of code synthesis and understanding tasks, Code Llama - Python is designed specifically to handle the Python programming language, and Code Llama - Instruct is intended to be safer to use for code assistant and generation applications. **Out-of-Scope Uses** Use in any manner that violates applicable laws or regulations (including trade compliance laws). Use in languages other than English. Use in any other way that is prohibited by the Acceptable Use Policy and Licensing Agreement for Code Llama and its variants. ## Hardware and Software **Training Factors** We used custom training libraries. The training and fine-tuning of the released models have been performed Meta’s Research Super Cluster. **Carbon Footprint** In aggregate, training all 12 Code Llama models required 1400K GPU hours of computation on hardware of type A100-80GB (TDP of 350-400W). Estimated total emissions were 228.55 tCO2eq, 100% of which were offset by Meta’s sustainability program. ## Evaluation Results See evaluations for the main models and detailed ablations in Section 3 and safety evaluations in Section 4 of the research paper. ## Ethical Considerations and Limitations Code Llama and its variants are a new technology that carries risks with use. Testing conducted to date has been in English, and has not covered, nor could it cover all scenarios. For these reasons, as with all LLMs, Code Llama’s potential outputs cannot be predicted in advance, and the model may in some instances produce inaccurate or objectionable responses to user prompts. Therefore, before deploying any applications of Code Llama, developers should perform safety testing and tuning tailored to their specific applications of the model. Please see the Responsible Use Guide available available at [https://ai.meta.com/llama/responsible-use-guide](https://ai.meta.com/llama/responsible-use-guide).
NeuNav/Reinforce-CartPole-v1
NeuNav
2024-01-30T16:20:07Z
0
0
null
[ "CartPole-v1", "reinforce", "reinforcement-learning", "custom-implementation", "deep-rl-class", "model-index", "region:us" ]
reinforcement-learning
2024-01-30T16:19:56Z
--- tags: - CartPole-v1 - reinforce - reinforcement-learning - custom-implementation - deep-rl-class model-index: - name: Reinforce-CartPole-v1 results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: CartPole-v1 type: CartPole-v1 metrics: - type: mean_reward value: 500.00 +/- 0.00 name: mean_reward verified: false --- # **Reinforce** Agent playing **CartPole-v1** This is a trained model of a **Reinforce** agent playing **CartPole-v1** . To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: https://huggingface.co/deep-rl-course/unit4/introduction
gg-insoore/detr-finetuned-cardamage-v2
gg-insoore
2024-01-30T16:17:16Z
181
0
transformers
[ "transformers", "safetensors", "detr", "object-detection", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
object-detection
2024-01-30T16:17:06Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
adalib/fate_flow-data-codeparrot-prefix
adalib
2024-01-30T15:56:43Z
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:codeparrot/codeparrot", "base_model:adapter:codeparrot/codeparrot", "region:us" ]
null
2024-01-30T14:11:21Z
--- library_name: peft base_model: codeparrot/codeparrot --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.7.1
Ashish1310/phi-1_5-finetuned-gsm8k
Ashish1310
2024-01-30T15:55:28Z
5
0
peft
[ "peft", "tensorboard", "safetensors", "phi", "generated_from_trainer", "custom_code", "base_model:microsoft/phi-1_5", "base_model:adapter:microsoft/phi-1_5", "license:mit", "region:us" ]
null
2024-01-30T15:26:16Z
--- license: mit library_name: peft tags: - generated_from_trainer base_model: microsoft/phi-1_5 model-index: - name: phi-1_5-finetuned-gsm8k results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # phi-1_5-finetuned-gsm8k This model is a fine-tuned version of [microsoft/phi-1_5](https://huggingface.co/microsoft/phi-1_5) on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 4 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - training_steps: 1000 ### Training results ### Framework versions - PEFT 0.8.1 - Transformers 4.37.2 - Pytorch 2.1.0+cu121 - Datasets 2.16.1 - Tokenizers 0.15.1
ostapeno/library-stablelm-10-experts-flan_3ep
ostapeno
2024-01-30T15:53:38Z
0
0
null
[ "region:us" ]
null
2024-01-29T22:25:17Z
Number of experts present in the library: 10 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | c2o10_5e | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/glue_sst2_2_0_0,natural_questions_open_1_0_0,duorc_SelfRC_generate_question_by_answer,lambada_1_0_0,kilt_tasks_hotpotqa_final_exam,glue_cola_2_0_0,ag_news_subset_1_0_0,cot_gsm8k_ii,kilt_tasks_hotpotqa_combining_facts,duorc_ParaphraseRC_generate_question,kilt_tasks_hotpotqa_straighforward_qa,cot_strategyqa,cot_ecqa_ii,duorc_ParaphraseRC_generate_question_by_answer,gigaword_1_2_0,kilt_tasks_hotpotqa_complex_question,glue_qnli_2_0_0,kilt_tasks_hotpotqa_formulate,para_crawl_enes,cosmos_qa_1_0_0,cot_esnli_ii,cot_qasc | lora | | c9o10_5e | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/anli_r3_0_1_0,paws_wiki_1_1_0,glue_mnli_2_0_0,glue_mrpc_2_0_0 | lora | | c6o10_5e | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/cos_e_v1_11_question_option_description_text,cos_e_v1_11_question_option_description_id,cos_e_v1_11_question_description_option_id,cos_e_v1_11_description_question_option_id | lora | | c1o10_5e | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/app_reviews_convert_to_star_rating,app_reviews_categorize_rating_using_review,cos_e_v1_11_question_description_option_text,math_dataset_algebra__linear_1d_1_0_0,dbpedia_14_pick_one_category_for_the_following_text,dbpedia_14_given_a_choice_of_categories_,app_reviews_convert_to_rating,dbpedia_14_given_list_what_category_does_the_paragraph_belong_to,glue_qqp_2_0_0,dream_baseline,dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to,qasc_qa_with_separated_facts_1,cot_creak_ii,cos_e_v1_11_description_question_option_text,qasc_qa_with_combined_facts_1,qasc_is_correct_1,cot_sensemaking_ii,qasc_is_correct_2,duorc_SelfRC_generate_question,definite_pronoun_resolution_1_1_0,glue_wnli_2_0_0,cot_strategyqa_ii | lora | | c8o10_5e | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/adversarial_qa_droberta_generate_question,cot_esnli,cot_gsm8k,cot_sensemaking,dream_generate_last_utterance,gem_dart_1_1_0,gem_common_gen_1_1_0,cot_creak,duorc_ParaphraseRC_title_generation,adversarial_qa_dbidaf_generate_question,fix_punct,duorc_SelfRC_title_generation,gem_web_nlg_en_1_1_0,app_reviews_generate_review,cot_ecqa,dream_generate_first_utterance,gem_e2e_nlg_1_1_0,adversarial_qa_dbert_generate_question | lora | | c0o10_5e | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/anli_r2_0_1_0,anli_r1_0_1_0,glue_stsb_2_0_0 | lora | | c4o10_5e | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/adversarial_qa_droberta_tell_what_it_is,adversarial_qa_droberta_based_on,adversarial_qa_dbert_answer_the_following_q,adversarial_qa_droberta_question_context_answer,adversarial_qa_droberta_answer_the_following_q,adversarial_qa_dbert_tell_what_it_is,adversarial_qa_dbert_based_on,adversarial_qa_dbidaf_based_on,adversarial_qa_dbert_question_context_answer,adversarial_qa_dbidaf_tell_what_it_is | lora | | c5o10_5e | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/adversarial_qa_dbidaf_question_context_answer,adversarial_qa_dbidaf_answer_the_following_q | lora | | c7o10_5e | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/cos_e_v1_11_rationale,cos_e_v1_11_generate_explanation_given_text,cos_e_v1_11_i_think,cos_e_v1_11_explain_why_human,cos_e_v1_11_aligned_with_common_sense | lora | | c3o10_5e | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/duorc_ParaphraseRC_movie_director,duorc_ParaphraseRC_answer_question,duorc_ParaphraseRC_decide_worth_it,duorc_SelfRC_question_answering,duorc_ParaphraseRC_extract_answer,duorc_SelfRC_answer_question,duorc_SelfRC_decide_worth_it,duorc_ParaphraseRC_question_answering,duorc_SelfRC_extract_answer,duorc_SelfRC_movie_director | lora | Last updated on: 2024-01-30 15:53:37+00:00
meiyun1995/Reinforce-CartPole
meiyun1995
2024-01-30T15:52:56Z
0
0
null
[ "CartPole-v1", "reinforce", "reinforcement-learning", "custom-implementation", "deep-rl-class", "model-index", "region:us" ]
reinforcement-learning
2024-01-30T15:52:47Z
--- tags: - CartPole-v1 - reinforce - reinforcement-learning - custom-implementation - deep-rl-class model-index: - name: Reinforce-CartPole results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: CartPole-v1 type: CartPole-v1 metrics: - type: mean_reward value: 477.10 +/- 68.70 name: mean_reward verified: false --- # **Reinforce** Agent playing **CartPole-v1** This is a trained model of a **Reinforce** agent playing **CartPole-v1** . To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: https://huggingface.co/deep-rl-course/unit4/introduction
Kavya2003/my-pet-dog-xyz
Kavya2003
2024-01-30T15:52:13Z
0
0
null
[ "NxtWave-GenAI-Webinar", "text-to-image", "stable-diffusion", "license:creativeml-openrail-m", "region:us" ]
text-to-image
2024-01-30T15:51:12Z
--- license: creativeml-openrail-m tags: - NxtWave-GenAI-Webinar - text-to-image - stable-diffusion --- ### My-Pet-Dog-xyz Dreambooth model trained by Kavya2003 following the "Build your own Gen AI model" session by NxtWave. Project Submission Code: GoX19932gAS Sample pictures of this concept: ![0](https://huggingface.co/Kavya2003/my-pet-dog-xyz/resolve/main/sample_images/xyz(2).jpg) ![1](https://huggingface.co/Kavya2003/my-pet-dog-xyz/resolve/main/sample_images/xyz(1).jpg)
Artefact2/Midnight-Rose-70B-v1.0-GGUF
Artefact2
2024-01-30T15:43:05Z
68
5
null
[ "gguf", "en", "license:llama2", "endpoints_compatible", "region:us" ]
null
2024-01-19T12:54:46Z
--- license: llama2 language: - en --- <img src="data:image/jpg;base64,/9j/4AAQSkZJRgABAQAAAQABAAD/2wBDAAsICAoIBwsKCQoNDAsNERwSEQ8PESIZGhQcKSQrKigkJyctMkA3LTA9MCcnOEw5PUNFSElIKzZPVU5GVEBHSEX/2wBDAQwNDREPESESEiFFLicuRUVFRUVFRUVFRUVFRUVFRUVFRUVFRUVFRUVFRUVFRUVFRUVFRUVFRUVFRUVFRUVFRUX/wAARCAGAA70DASIAAhEBAxEB/8QAGwABAAMBAQEBAAAAAAAAAAAAAAECAwQFBgf/xAA+EAACAQMDAwICCAUDAwUAAwAAAQIDBBESITEFQVETYSJxBhQyQoGRobEjUmLB0RUz4RYk8UNTcpLwB4LC/8QAGgEBAQEBAQEBAAAAAAAAAAAAAAECAwQFBv/EACoRAQEAAgICAwABBQABBQAAAAABAhEDIRIxBEFRYQUTIjJxQhQjkaGx/9oADAMBAAIRAxEAPwD8jABUASQAwCewAgkAASiABOPwIxgsmMeN0ATLLDXsUx4IyFXcccbomM2vf2KqXkthPjkDWElIs2c+8X4ZpGontLnyRdtca1jlmc6c4b7o0pvE0b14v4Yy4e6ZNta3GVGvCdJ06q+PtJ9/mdVpYupJSVVKTeIxZEbGnVt4NrTNrlGFGvU6fXUaizBPO37onv0ute1bq2nSvK1OacZRer8C9K5cEqVdaqT7f4On61Gpd060nqjhwb8rsSlQua9XMHGLjsvDG/01+Ou3rztYp5de3xtJbyiv7opewpXE9UGpao5Ul3PLpXVSzq4i3KPeLOyNSE/4tBLnMocf+GT011Xn1aChrym2/stdn7nPg9K5nGrKGjh92t8nJVp53Wz/AHNyudjAlLKyuSYpOLzyVTaexpldvK35Ky4NNSqezKSWERVUWzsVRL2CIwSlsME9gKtAkARgEkFDBK2BbGQIIJABENAlMCYPDJbeSrZblAS8tFoJvYhfZLwkRVlFpnTbx+L4uO5zueX7nRb3MqE86YzWMYkSrJNup0pQnHK2e6fZno2r0x/E8qlcTnFQlLMVLKXg9CjL+GuftGa6TX0+4+jtJuhTqY2TcsmXX/o5WsHc9SsM1IVtTqU/vRUudPlf5Nfo7e0pWVOhByp1OHDlSflN/sfUTuKdWhpk1GEo7Sclg4+q7e4/KJwtn0WlVhSqRuFWcZzTzGUMZW3n/ByUaVW8zG3bcY74k8YPtPpLbQtVb3NGPp+pKUZVdOYxn2ylt5/M+YtKztepS9NwzLDjp4bWcPf8zcrHjHDRoN1f42tOONl2Po+uu1s3Sh0elOhTlTfrVHvKe3Gp+V4PKpVIVrmbk/iknnfO/Y+voWtsun2tStb16spQVRasJU1y3LPZEt7WST0j6O9FXT7OheXLUbi6y3lcLlROD6Q0HG+uJSwnjVh/JH1d9dU4dLpzor6y9cYxalhJ+77I+H+k3WJ39WnB1HKVNYm4tac90scok7u19PnbjGpnFUi1h+d0bVp7+5zzlhnWOWSuJYk0tvJlJvB0JzdGUnmNLdZa2b8fMwlLKwsGmGbjJJFNLybJuXL4Jik034BJtzyi5TSI0uU/ZHQliLk+4jKMKEm+WNmnJUTk9jPS8m0e7KuW7KyrgS2SRKe+WQ92ETF7GtDGvMjEtF7+wWLVJepN+DBrc1UdmZtbgpjYjBJPBWUYIJyFuAe5HBL2IxsBHJoto+5QvjYKzaGCd28JbnZbUEvjb3ISbUoW+HqqLfx4OypVjThqk/w8mdSoqeMLVJ8RRk6qoy1y/iXD4XaBlv00cI016939r7lH/JnGlVvnOrKaWNty1vbSrzdW4bfs+52fDTg1BJLOcLyTel1ty06saFX0E1pSy2+7IhWpq6qSa17fDjjJRQdeTcktuWzBzVNYgXSWr16s5LTN4iuy7nM25P8AsWSc3lv8SzxFbclZ9pjS0x1N4f7FW8vTHdj4prxHyS5KK0w/+3kojSqcucv2IlJyeWyBnARJUZyWitgK4yTjDLrgowIYwSAIwTjIGrwBD2eAl5IyMgWcuyKgAQCQERgEgCBgkAQSgEBCJAAAAAAAAAAAAASmQAL7P2fkhrHP5lcllIKhr8RnBfClvHZ+CrXnZgSpZWGQ1j5ENNEZYF41XH3Xg64XacXHSnlY37HCE8PKFm1lse5RuKUYRjGXzjIi8hCpSTeH3TPKjUzs+TTXJxUc7LhGPHvbp5biilKjJPlfud9re06UZLGlyzu9/wADjk8pJmTg0vh/Itm2N69OqpSjLM3zj8ytScqU41IYimuEZ0auFplx4ZtNRnRwu3AX36a0/wDu05UWoV4rLi3yjnllSaktL8PY0co0qUZQjiSxhryRVrOtiFeKjU+7NdyRawlDO65/czwapvVpns/JE4qS8M2xpknhkyeSUuz2aKsIIs+CFyWktgKplsZRQut0CISDW5LeGMZQEdipZ7FSiS0diIomS2Io1hlWWTysFXyVAABAtDnBUlAacJomJGchbMjS6xk1TMluy/AHRQ3mezSoqEYxUlKWrdxlmJ4lvLTUR69vUThHCec7vOzM1vF9N0erV9OVvThCTl/Ehr+40t/zXY9y2uqcunVKUY6ZyTi4PZxljjP7HkfRyNGpPVKm3UhunF/F815PXvqKivrdOnGSl/uxg/tRX3l7rwcb7eielrzqtF0p0acfTVxhy9RrQoJLnPc+YdK0p9duqtGlCVrGTVGK3i33a84+I6Op9EuOo1qFtTqU6dGSc3VqT+088RX5fLJ87Vq14V6UaadKVPMaeY8Jbd+O4npPT1bWtS6f1qr/AA0qkpJ0XOPGedj6Snd1Pq9NSuZVa8V8epY0/wD7B8ZCpddToSoOrqqWsHUpqSb1JcrJ7/TI1brpHqVptua0qUZbtds+/Io73dq+oVKKq03GG9SpP7Ckvu++T4+7gnUefhcnu1ukfX0Z06HwQjCOIuMI5y/wXC+fJ8lfzcqspacZ7FhXl3VNxjGprhJTztGWWsPuuxxVDpnOLTymn2aOWby2dI45Xakpy0adT05zjO2TPJeRm+TbktrW+OPcKTK4LRaSCxMpbJN7FZtPOnLSKvcrxlEXaY7iokI8kyWwGWCeBgLncrI/hXuyE9sBvLyyANYPMWZvktB4iyreQqCCeSCoYyyVsQtmXis5z+AFcBlm0kU92QTFZeWWeXtEqk2zeEARNKmopeXyzVyaemn9ru+0Sr2T3xFcyM251IaaUWoZw2u5GvSYpzq+lQeqcuZs1oUHbzdSssOPCfctZ2/o3FKUn8Tz+xW5bpVqkM6oOWcPcn8L/LS9q4jSlRnJKWXhGHr4Tlw328szuKym4/Dp0rCWTDLkxIXLto6stGnOF7dytOm6knjhbtvsRhLPctSc23GDwnyaZ9+yfwTcU8siVLQlKo9393uaSnCjLMPjqd2zCUnOTcm233ZCrTqOe3EfCK8EZIyVEtkckqOfZF0kuAKqPklFiGAbKhshsCSMkABkMAIAAAAAAAAAAAAAAAAgkAAAAAAAAAAAAAAAAAAng0U01hmYCtHt8vcq4Z4/IhSa+RdYfGzAzaBo8PaSw/JWUGvkBUvCo488GYA6cqSygYKWODWE88kXaZRUvmRCbpvD4L7IhpSQVfWpLSkscl7vFSnGfC8I5nFx+RKllYfHgaNkZtLRUWV7lmtK8x8lpZqNY2S3z4KSko4lHeL5TAlxUluZuLXJq8OGYbr9UQmn7oqWMuGWbyTOGN1wNnBYW/kIzfJeHBVl44wCIlyiy4Kvkt2CqvkglrcRg5tKKbb7IqJixJ7HbS6NeVIavT0r3ZnU6dcQeHBmdxrwy16cie5L3JnTlTeJLDKpmmQABAkJZaRacdMmucAQnuWTKkoDotY06txGFWsqMHn42spbFE9zPBZPHcml26KUlqR6drNKm8nj03iaPVsoylS14elS5xsStYvoul1ZRqUnFtbPddtz6CV3NZnnVq+0n95e/v7nzXTqsVVprxk9u8qQoQS0ynGclFKCy9zjZ29GN6cXWrypd0YxrwmoUfijKUtMksY3+eFz4PDt4QoqNepXcquMJSw8fqR1vqlS+UKMKnw5c54XMm+/yR5EZKrS3cUuMY/uWTpm5Tb36coTqp0a6p66eioo088vfvwfQWklC1p2kYtxoY+3HGr3PzyUfTllPGO8dj7HpHWPU6JUpVKWqvReYSSbc3/42Ys0sy29yvUdb6pN6cqWnEYqK7+D5Lqm1zPtsz6ONzFwpTpxeFNS0uOMex83e1Y3fUlCT9OMn8Un2Xd/l2EMr08avHSo+5yyeGdt9KE6050FP6vqehyeXjtk4Kn2tjrHCqylhlWyZQcX8Sw3vuUyVlaRGcIh8FWwJT3HLKjIVrBZkJc4KKeCNbYNpk8FSG8jIRHcnBGRkolbIgnJAAh8k5I7gCU8EMlbMBjG75JUXJ+xKjl5ZrHCTIulUlE0TSi2/wDyZ6sSTfBdJycZNfDlbEWEISuq0aecJ/obfFZUFOnNNzk0012RWrJUrmThHS1tgxq1fUilN7ReUQ9Np3HqNSg2nF9zllUw3h5fllJTctlsvCIS/E1pLUpOW7/MnKWyEnnbuTCpGmntmXbIGiopR1VXpXjuUnV1bRWmK7LuZyqSnLMnlkZBb+DIJw2W08NceX3CKJFlHHPJbaK2/MIACG8ENgTq8blW/JDf4AA9yAAiQAAAAAAAAAAAAAAAQSAAAAEEkEgAAAAAAAAAAAAAAAAAAAAAFlPs90Xi9vh3XhmQTwF20cVLjZ+DNxa5Lqaf2vzLOW2+68gYhbGjhneJRoC8ank1Tzl9jmLRk48AldBEYKUhComuN/cvGUU98kaUy6Sa+0mTRcGnGpvF7p+C84rLXYxlTxvH8gVtbQ1xlGKcpRTkku6XP6FJQa+LGnO69yKcks74eDSVOpUpQS5jnHuRfcZqWrbh+CGsccBxkktS+TJzjkrLJ4JWMF3HPBTOCoPGScojLYyVGtCkq1VRTxnufR2lC0sIKcsOXlnzVJuLyuTWVSUvtSb+bOeUteni1j3Z2+pfWLeWUmsHn3XVqcc6FlniNmcmZnHHXLnumleu69RyaXyMiAdnit3dpBBOQh3JzuR8ycgSERkAXLwai91lHVbWttdW8Gq06NVZUtUdUG8/mtsFbvptzZJSqQ1U3xUg8xf4k2umWU38PHyPoOkR02M1KeHLeMX3Z85BNyxz8j2LTqd1Yzi9UYLGyUNS/EmUbxunq2tGp9ejKnCTyn8KWcHsX/VLeVvQjCslXoVYyUZrD25PLtrmpf05XE3TguIThJwcpey/8Hq3/SKkrTT1pQeIfDfUd/Tf8s/88fI5ffbr9dPlet3NO5vncU6XpuW84x4cvJw2lldXNxRo0KT1Vntq427/ACNPqlf659Uj8VRy0aecv/B9VYdKo9Icpep6lx9mVTwvCNWyRiTdfGXdGpQu6tGoviozcZafKZ9X0e+p2fQqFOGiVxOqqmE8vnj8kef9IenTuLmV9bR1qazUiuc8ZRwdIpepV9WpT1UaPxTzsn7ZHubJuXT7CtfO7bkqThNbNKWrVvsfN9RoKNXVOaalvhLOEezO/q3lbmVZuOIU6ccRhtxhfkeP1OnWVROrOnSkk805VFqXtgmO2srNOGhVowVShcJ+hLOZQXxR8frg8uTae/J1TSbeMnJP7R1jjVWyvc6bOxuL+uqNtTdSb7Lhe7fYvf2qttMY4bh8E5ReU5eRtHI2QHwQUGyGySoRIyAgJIJIYVADGQi2duAn7FcjIF214K534K5JScmRUv4uxrTjhZwIU0uTdYe2CWtSMVFZy+Ck5rOy2NKy2a/Yh0k4RYLEwpKSUm+XwXrNP4Vslu2ZuelY8GLm5MCXJQ2W7Kby3YaxyTnCKyjGA34/MhyKt5AnPggAILcukvGWQltuTnbC4CrYS3e7IcssjtvsiHLwBL25K6vwQyQAbA7AIgEgAAQBIIJAAAAAAAAAAAAAAAAAAAAAAAAAAAAAQBIAAAAAAAAAAAAAAAIJAAtr4xsyU1LaWz8mZKeGFWlBr3RUvF74/cmpoUIpRal3fkDNMvGp5MwB1KeSTnjLCZdTzx+Q0u1pRzxyTCvOnheOxKeSGs/Mi/8AF5y1WzzypbZI0aYRcnlPl+Ck25NOSy+7L606UoLv2YENafstNexWS1IODg0/JPfw/ARlwwXaz8yj2NI1hwWMlUY1tk06TOSLzl2MyckMrGWWwABkAJAJDBOAALKOEpST0vK2KnVGS+pTUVvKcU9lvhAjo6ZDXRqb8SXPyPTtq87eU4SxKjODUqcllSPM6XPRKSnOMaUuU5LKfZ4O+f2MqWpexitxw39CnTnGpbwdOPDim3h/ieh0zRXpKctktt++DDHq5jjVnbBs7Gt0+UYV/hqxjp0KWdC5x4yZt1NOvFx3PL109qlC2VtUUqSVLnOrB6Np1y2VlO3q0ldU6unVqqaZ4XZ+UfLVbmfpQpbbZe75KwqTynPd8Js4yX3t9WzismPi9KhcK26zK5jQpRjxogto/JHVO4U5Sys6t2eLKrL1ZPVnc66NSVSkpx3jjdpmpu+3j+Rhhh3g6frDjJpb/Fnd8HFa2cKnWYzuElaznqmtWN/BhO6dN/1Mx1es3qi5ya+HD3i/J0k128nu6fpNO96R062/7WVPW1hQpfaZ8JfW0aupSrcSk90ts+WUjXq06UKbynGWcNbomvcTlTb+2vEuDncrvp9Hi4OPw3n7ctHpcZJ+pWwktlBZz+ZwdVtqdC4gqEZqMoL7TzvlnrUeoemszor4Uvs7foc11UlUdXXJNSmtl4NTLLy7Z5ODh/t7w9tqfUHTsXQt4KlThFRSjy33k33Zxzip0pwSTUlyzVRgqWtceCs28tRi2dXzXjSjpbT5WxXgvUeaknzvyRg0yqQWaKhAAAAAFCpYhoCAC0I/EshExg2axSivcZxsiG1H3ZlpbZLMiYKU8NPCyUjD1E5SePYs5enBLgKtL/cltsZzqvSVlVcv8lcbbsaNq7zZKxEsuNiknuVEN5ZDZDeSMhkYAAAJE7IBjJOccfmyucjIVOfxZDeQAgAABBIAAAAQSAIJAAAAAAAAAAAAAAAAAAgkAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAgAkASpeSCAL4T4K4wETnPIVAzgYARdT8m0Pi7nKXjNx4Cyuj03jLf5FZrDTT3IjNaXlkSecNEVvO4hVhTjKGmUNs+SlT+LNYWH7GeMkqThNSGjaM5XuR8zZRdSjOWY7PsYx4ELENEFiDTKCSCQAAAHsdO6RbXlP1KnUIQXeGh5T8bnjnX0+r6dwoviexL66WPo/8Aprp9ShH0LipKTX2lJPH4Hh33S6llOayqlODw5Lt80ejSqToT1QbUvKJubxVaTjNZnvlp/aT5TRieUrXVfPs6+nRoVpVKFxVdL1EvTnjKUvdHNWp+nUcVuuzfg9/on0etuquMo3FWE4qMpRnTTj77pmrZIzJ2459Kjbwq+pcRlUWNCgsqS758GEINZ01En+R7nWeiXnS6cvq1xGpRlzGfwz38J8nJR6ZTdCKlNOot5Si/0Meck7r08fx8+W3xnpjYS13CjJ7rx3N7i7U7iacnJ6uX3Jq2dK0iq1OLS7PWZqzUo+p6j+J5aayZuWN7ejDg5uO+OLGvPOM5w+Gu5NOqoR1L7S8pFaleM4ypVYYnCWFjujm9enSbT3fsXW4tzmOW9tZXTitK+0+7NKbuKdOVSgp6U/i0v+xwwlOvVlKK+JJtex61rb06NPMVqqS5nLfHsi2zGOUwz58uvT0alKzuPTk4SnpW7hLaXzZ59SjOzdRqP8FvMZJ52Hryi3TcpY8Q2wXgqNahOMnKE0vhqqWWzluz29n9rH/wmqj6x9beYxzUS28GlCtTmnRqNU5NrEm9jyKV7Ws62qKi5Re+eGb3PUfrkFmhSjL+eK3NXD89OWPyJrdvf/66rmMYSkqbVR4w3DdfmcEm28R5eFuVhWqaNGtqC7Z2Ma1bE9K3Xf3N4zTjy8ks29Gdpc0paK1OdJZ+1OOxalYqtOE69dvbLhGWF+PsRDrVednC3unOtRhLVCL3eMY/FJlZ1s13KMnKLisaeTOVy9NcGHFryy7OoWcqs4uk6Tx42OK5s1b28JuqpVG8Sit1HxudcLrEWuE+7L64aVNJJz5TWxMblj07cnHxctuX3W1p9HKMFQqdWvqVrGqlJUk8zcXw/bJr1HpnQKNFStrmvJqTTa3z+aPOqap1HiT2eM5FSEqkYrXt7nWS3vb5ucmN08ypFRliOce5Q3uoqNdpNNJdjBm3IBAAMgkASti6ZRIsv0JViXLHBZLDyyHBN5TwkMuey2RGk60otIrLMn8TJUcEdwKtYQ4W5MppIycmyomU+yKcgBAABAkgAGwAAAAAAAAABBIAAgkAAQSABBIAAdgAAAAAAAAAAAAACCSCQAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAQBJBJAEgEATkAgCQAAyXT2KEoK0TJM8kqQF1lZwyE8JruEw1lBUrGn3KtYGdkidWVgIgDAKgAjSnQq1cenTnLfHwxb3AzRZbPK2ZtOzuKeddGpHHmL2MklnGUvmwOyj1Fwg41Kev+rVhnVa2l51OFSrb28vShzNvC+WfJxUKFGcZKVSLeNtMhT6hd2sfRhXn6cX9hv4V+Bm/ws/l6br0+n3Lo1vTr5p6HU0J+k/bPPg1X/Y0fXs5yp1IpP1ae2d/2PEU3W1S5ly8s77eUVbQgq8pZmnKDh8Mc8bmMsXbDOassK15d3txGpXrzqOXw6pvZHXGn9WpOcqm3MY55XY5brT6+mC+CO0UuyMo15Sl6VTLi46XnlGbN+np4M/CdvSoXEKuiE5xSb3bbzj5G9KpSdR0JcptJvwePOjRhhRnJ+yWCdU1JST+NvhckuO3bHnywvbe4tY3nUm1OUYxhjMe7X/k47uxVBLDcm2/iOyjWjazl67xKMdl5bOerXd5JRgm23sjWO5/xw5rhZbf9qmycI2kkk/Ucvib8GquZQlFY2WU0+5nTsbuNRKNCcs8qKz+x2f6VdzlGPoNN9nJIWRMOXKYzXWnG6sksNtJkQrOUXh8vg9X/p/qUqcnHp2N9pKqn+hw3PT7y1/37arTS7uG35jprzt+3HWt51YN04Sm48qKzsZVqNa2loqRlB4Tw+2T7f6J2k4UpVWt626lqWyH0g+j0ZwrXt1f0KEP6o/pnuyy/Tz5zvcfB+rLeOce56tDo0b/AKUriym5XFL/AHaLeX8zy6/oxrr0ZSnFLdtcs6bfqlxY11VtVCmvCXK8M3Z+OW/0hUWmlDDbjHTj9ysf4Uu8WnsyLq6hXqzqUqcoSb1Zb/8A2TGncOOVUWpZz8ieLtjy99uyr6U4pwjFyxu12M9TTWXx5M4TUt05Y9zacJ1KcpRjGSgsteSenW25dxvTi01q7+RVqRoxbknjwcNJxnFJbS+ZSpqhJxnybjx27Z1Ja5yljGXnBQ9u16RbTt1VvLl0nJZjCOMnnXdrToSbpVo1Ids7S/ISypZY5CSAVAlIJF0t9hVkMYwG+yGSEtzKrJPGGXhEhbETqqK25DROWGYyn4Kyk5MgrNqGwSAygAgCQAAAAAAAAAAAAAAAAQSAAAAAAAAAAAAAYAAAAAO4AAAAAABBIAAAAAAAAAAAQSAAAAAAAQSAAAAEAkAAAAIJAAAAAAA7gAAAACJIGQJBGQBZPBOrJXIKqzIGRkgnIILLgAd3S7+dhdwmpzVJv+JGL+0vkcS4BfaS6r7a8rUbmhFW1aPqT+KEs91+x4Ver9Yh6dw9STf3U8fI8qlXnRknFtY8HZbxldRm4RlpjzlpJGJjpu5bZ3vTJ2jTypKSUklJN4/A48N+/wAz0ouWZQhFSWM6orOPmc2r06mqUU87P3NMsfSmpKOluT4S3yei7OFCk6Uqn/dpapxctopfd92dnSalJVXVw9cHiC43857HLfZubpyo0oJy7wb/AHZm5d6dMeO3HyYSnUlFS1aZd2iyam/i3ku5VqtCa9ZOce+Gn+xte21GOmdrU1U2lnPKY6rUuWP0pCEpT040pd2bwUKU04NPHcxtaHqt0oL1Jya4f2T3bTpDofw4uVOct5SgsP8AAldMc79MrTpVW+k/Xjopy8r4n/g9q06PRtqWj00t+VHL/Urbwla3EI1rmUo1HiE2kt/5ZL9mexO7hRhvJbENd7qlKyi4Ll48siFKnOo4/HSqw7Rlh48+6OVfSGhCrplLZrZnmXfXIyrr0niaeYPw/HyZNG4+lV7O2TVw9cEvtJbr5r/BSv1W2UNUZxqJ9vK9j42r9Lq9ZOFWhGnJZxUi+PwPIl1CrKtKaxFT3cY8MsxrNzj627uLKrHVGjRU4yym4bP5iMqHULadKh6dGbW8Gk0j493lTU3q2fZiF7UpVY1IvEo/qa8WfOI6l0u56fWkq9LCb2lH7LOetPEtGPsrT8z6H/XFVpqlcx9Sk+77Hj9StFRqetReqjPvn7LLP5c8pr046TxJe+x3xs4uVOVTdSfCXY4qEVOTUtlzlHuyrUaNG1+JPdJruZztnUej43Hjlu5fSfRt6clGnCGnG7bN6cI+luksS+F45yYYjLVUjJJPbBySutGmMXF6ZasLhHLx2+h/cxwu7FqnS4urKVNum48vsjguY06NwnKs67TTcdOMrxk3r3FScnVqSbXhcIw6i6VWqq1BqUZxWV3TXlHXHf3Xg57x/wDhiyu76td3Eq02ot8RisJLwc7bcsvn3PpOn9ItKljC4dSVOrBZm5uOPyZ49zJ16k5t68cNrlG5Z9PLZftyYIGQVlaPJZLBRFs4I1E8shtRKueODNvLBtpKo2ZtkAIEkAIAAAAAIBIAgkAAAAAAAAAAAAAAAdgAAAAAEEgCCQAIJAAAAAAAAAAAAAgAAAAAAAAAAAAAAAAAAAAAAD1ra2pdVpSw1TuoLfxNeTz7i2q21V06sXGS/UWtzO0uYVocxfHld0fa1LW2v7aMpQVSnNaovwcMs7x3+H0+Hhx+VhddZT/7fCYB7l90CdJOdtJzj/JLn8+54s4OMmpJprs+x1xzmXp4+Xgz4brOKgEGnBIAAEEgAAAAAAAACCSCQAAAAABkkgFFsklSQLLYkqb0Iww3N4fYCPSzwjW3q1ZU5UacZNJ5aRjWn2T29ha1pUK8ZweH+5Lv6ax8fKeXp6fSqV9Rv6dS3pTW+JNrC0vk6er2jddSlHLly0/BvadepRi00qdThNrKTOfqPX6lzb07abhVjSlmNTCU/GNuUcJllcu49vLw8WOHlhlty0Zxoxnp2WcFnWbpPTp+TXY417bolSlFYxlG7Ns453GaiJfHJ5Wn5HZZ0LapRrSuZyU4r+HFczfg48LZt8EqeiEquMT+zD2z3/I1pyt17fR9IUdLnphCo2klBbG3/UUKN1OnWWuOtxjOPZHzdO+nSxoeNKwvmZUbarXp1ZQi5Rhhya3wPH9Z8/qPa6h1iNeTj917Sx+jOOp1SvUoaZVG2u/k89RyiJJxafZ7GtMXKtJVpvuVU21u+N0Q1sQXSbTKeubcuX3Ixp+RXuWfGewEMqyYc6Xw+GSlmTXcIQl918PY1pV9MXCe9OW0omMlpZXO7C7TJSozcVL8fKNbW7lSnJOMajmtPxdjNaZJas7ePBVpRr/DNNJ7SM3trG3G7jbtgtGLk1jZZxlkqMV8Uk5R5xF8kSm6q1YUYriK7GXeSfbqq0M0ZYq05pLhHPbKhC3uKVxF65x+ColnDMnVlGO0ouPginP1HGKjl+4ks9ryXDPWppnVq1VTjScs04vKS4yRB1JpxT27tnXGzdWhUnConOPMGt2jlp1VFNS2Nzt5spYh289ajBa2+FFb/kVqUqlJL1KcoZ41RaPe6era2i6larD1ZLjPC8Gl/wBUhG3/AIUozcfsxnHMc/iTd30amu3zSeEQ3ktNucnLbLedipWUMAgIAAAQTggACQAAIAkAgCQAAAwAAAAAAAAAAAAAEASQCQAAAAEASAAAAAAEASAAABAEgBAQSAAAAAAAAAAAAAAAAAAAAA7rHqNW1klCo4pfkcIJcZZqunHyZceXli+wt+r07hKFZKEn3+6zO96bRuMqccPtJco+Xp15U1jmPhnr2PWVGKpXDen7sn2PNlxXHvF9fj+ZhzTw5XlXVB29xOk2paXyu5idPUHm/rPOfi2ZzHpx7j5HJJM7IAArmEEkASAAAAAAAAQCQAIAEgG9GhGqm/VjFrs0xbprHG5XUYA0q05Uqjg8PG+U9mZhLLLqhJBJUSjRRk2oxTb8IyPVtqVH0Y1EnKUlv7AcqsKrhqm4w9nyVdrOljVFxb2UpbJHt29rVruMktSj9lv+5PV6FGjYem2pV9Skn3f/AAZ21p5denWuZxX8DVjGuntr/wCTmlZ1qf8AuR0rOM5IVeUZ6c5S/Q1neynDTJJ+7W5UVhFwW72Jz77Er4o5X6kKD3I3NjrqC2WWRVqKpF4+7L+xSpTabfKKU3z7liZW/bWKykff/Qayj9Qua0op+pU07+Ev+T4SjH4c9lg/UvozSVv0O2jjDlHW/m9yZelw9vi/pJ0V9O6jUlRg/q8lrWPu5f8An9zwprVBrufrl5bRr3tspRUlOFSMk+6wv+D84690uPTuo1adu3Kh9pd9H9Lft/dDGmWP3Hkp5S90AvsxzyngdzTmpjlhTTjun80XxhGChKUlGCzJvGF3CxVz35NITzVT8o+v6Z020o0fR/hOpBJznKmpOUnzz2R4nW+l/UriNxRUVRlPS0vuv/DMTOW6dsuHLHHbzqhlgvN5wVzybcVqTxNe5p6Kqa5Y054yYJ4aZ0xnKpQnRjnLecLuRfpzxeiqsvZPsdXw0J7Ykn2NYdOVKi6tZ5fZHPWcEt9/GCXtvHeMUrRTepQUURu2uOMbFYz1YUpPBopJJ7rC7hZZR3FWk04Taku5i608TWVibzJY5I3nLjd9kddCliDXpucnznsX0xld1SlXpKCc3h+EUq1ZXM1CnGT32S5Za5t5acqi17or065Vpd06k1mKe+/AZddL6P31aClohBviM5YZV9B6jFtTtajSWVKOGj6J/SS3p7Wtu5vtOWx4/UfpJe3EJUlUjTTTi1TWP1MS5VuzCfbxKkJUqkoSWJReGigBtzD6LpP0bVzCNa8nKEJLKpx2bXuz55NJptZXg77nrV5cRcPUdOm/uw22+ZjOZXrF34bx43efa/XHZ07z6vY0oxp0lpc023KXfc8skg1Jqac88vLK1IIJKwAAAQSABZItSpTq1I06cXKcuEj6Wx6bRsKaqV9Mq38z4j8jjy804/8Ar2fF+Jn8i9dT9ePbdIuLhapL0oeZ8/kdFXpNCjB6pzlLzsj0rjqMVlQWfdnmVq8qrbkzljnyZXd6fRy+P8Xhx1P8q8uvTjSquMZZXlmZ03TXwrCzu8nMerH0+NySTKyAAK5gAAAACASAIBJAEgACCQAAAAAAAAAAAAgkAACCQAAAAAAAAAAAAAAAAAAAAAAAAAAAtJ5Udt8c+SpMpOTWXnCwQFAAEAABBPcAAAQBIIJAEEgAAAIJTaeVsAFS5uWMvOCAEm3hbsHsR223S7y7gp0becoPiXCOI9XonVJ2VyqUqmmhU2ab2i/IpHXS+itacE6txCE391LVg561jc9JqxlUxWoLvF7fkfSTuPTpNZcm+HHk8246vWSUatOnUqU/v1I//wCeDEtrVkcK+kc4wahQSfbfg8ytXrVpylOblOpu37CMKbus1Yy9JSzKNPnHselcztaydza2UbShD4U5TcnI16T28eMdvASwzWb1ZcFz47FUnGag+TSLKrLxk1VRfJ+5CSS8FJRy9iaamdizy4VJdlF/4Oen4N5NulNY8ZMIrcqO+3aqU1Rin6lSpFfM/U7OcaVCMcpRhHGW+Ej826JSVTqkHJYjCWce6PrL61r3kYQVR/V1vOmvvPtn2OeV7duPHca1+uSvr+Tt6zpWVOPpurHZz7vS+y2W/scF/wBZsKfpQi4yUW04pfda3X7HJTsal51GlYzqTp0tLlPS8ZLV+h29LTbQg3Wg260+0Uvfvnt8xj433WspljZ4zb56sqc69SVGDhScm4xb4Ri/9z8DuuoaaNu8YcqK/dnC/wDdXyOk9OGU7Gtj1vo/Z069d1ajeIPjHs9/zPKZ7n0YqabitT/nh+3/AJM5el4/9nppfVL6cZQ1JpS1QnznycH0muIRs3BPPq1FpzzsejTsqFO/uK9aq5SlFOFNbKK43f8A+7nyHWL/AOv9QlKH+1D4Ka9vJyxm69PJnrHTmluyi5Zd92Ujwd3jPJ0W9WVKopw3kvJzG9tLTXpvjEl3JSL3NzXuJ4qSb8LGDN0sQblJLwjtvq6rVW4Rj8L0ua5kzma1NQ4xvkT0uu9MoY4xv5JcJVNSysRWRKTjLGEmu6RpF04xc98/y55GzxYUtpp5x7o6lNasvdfM5k22WUU/YExtdruNFBY+KXZS7Hn1KlJ0nCENMnLMnnZrwb+j6tFJTerPjOCJdNqRTfqQx5JuN/28r6jjUnH7La+TIe7BBXJIAAgkACCQO4EEkEgAQSAOm3tXcRbVSnBJ4eqW/wCRyk5aeVyK1jZL3Nvet3QsYfwW6lVreo1+xnVuJ1HmUm/meVC4qR+9leGX+uSx9hZOM4pvf2+h/wCs3jMJ1Px2uRnKSSbbOOVxUl3wvYyy2bmLhl8j8Xqz9Sba47FADby27u6AAIAAB3AAEEgAAAABBIAgkgCR3AAdwAAAAAAACCSAJACAhEgAAAAAAAAAAMN8Bxa5QAAAAAAAAAAAAAAAAAAAAB3AgEgB3IAA1pU41G4uahL7urh/j2IqUalJ/wASEo/NGZ6VGtOVCOW2sYeTOVsd+PDHPq9V5wOutQjLeC0y8dmcjTTaawyy7Yz47hewAFcwEEgDqt683BR9SS0cYfBylqb0zTA9FaKyca0NeeJY+JHHcWsqDeU9Plo3pqdScYwxl932Nr2jOMo0XVdWb3aS7BXNb3tSgkozkkvfJpVup1U5VHmXnyY1abppZjhPgmjRndVo06e85PCQGeuWpv7OecHRK4qVKVOhFvTHiMT0q3SKfS4Sq13G4ko7ReyT/uU6P0+rVrfW6jcNPxQa2y/PyJvpdX082DdOTlFtN84fJRwcpuSZ6/UrulWpSpXFnTp3kJJxr0tlUX9SPMUWlkSrMd1WTlhJ/mW1be5E+xGPcuy4/iKiaWX3RSC2+ZrUTdJ/0mUN5RRUd9vUlaPXBfFnZrk+66FVqXfT4Vq0NEnlY+TPj7Czq391St6GNc3jVLiK8v2P0mVjR6fThRt68KsIpJY5OWWnpx36jgrWynNSS0tctdznvKSo2Vb044nJaYpd5S2/ueqo5PNv6VxfSjStkoU9813wuz0ru+3jkw3u60+O6lDVSo1IL+FShGjnzLGp/ujymm5qS4TPqPpFb07PptvbUliNOtJLPL+HOX+Z41naevZ9Qqf+zTjJfPUv8M6y9PPlj/k4G9/wPT6DPRdtrlL90zypHf0WWL7HlI1l6Zw/2jW4ur67rVqNKtJRqt01FPGVn7L9meLVoulc6Gmmucnv9RoVLScKlulpk1HblP5nhy+KvUk3nG2fJjFrkVltEqtkkWlu/kV7vPJ0ckLdkxfxNER5ZGcSYHXBNxi1xkvCKi3N/FJ/oZKo4qON4vlG1OLlUSytL3yzNdsNIlSU5amVo04RruFR8r4X7m22Gnx7nPVp7488MjeUk7aV4xlP7Ol9zJU5PYpUrtzSayorDfk1jNY+Fp+w9M7mVbxpujlRllN8ozvLjTQdNN5l29iYV2lhJfizGs3VrJvDWOcCRrLLWOo4wejoinmnDCKfV6dVvUnGXlGnm04AXnB05OMuUVCAAAAAAAAHYAAQCQBAJAAgkACCSAJAAAAAAAAAAAAAAAAA7AAAAAIJAAAAAAAAAAAACCQAAAAAAAAN7ZJywen/AKa61LKR48JuMk0fUdKvIVKSjLkzldR145Mrqvmq9vOhNxkvxMj7C/sad1BtLc8Gp0mpFvHAmcrWfDlPTzQdU7CrHtkylb1I8xNbcrjZ7ZAlxa5WCAyAAAAAAAAAAACCQIJBAEgACDvsk6lOUUm3HsjhNKNadvUVSlJxku5MpudOvFnMM9309uhY68Sq/CuyK9R6bGdLXSWKkVx/Mjnj1l1WvXjuu8eDpj1e304lJ/8A1Z5LOSXb7eOfxOTC4b9/rwCTa5lSnXlKjlRe+Gu5geyXc2+Dlj45Wb2kABkAAHd0+hO4rQUnKNLO8kj6+3sKNtTlFR1uf2pS5Z8PO5rTwnUkkuFHZI6J9VvqtH053VRx4xkzZa3LI9Dq19Tq1PSt4pRWzk1+xXpFxRtHWr1k8QSxhZ3fY86pcepGmpPeENJm6spQUMvSnnBrX0m+9vT6x1NXz/h/YznL7nJZ9WurJaack44xpkspHK+MIiMdUsDX0m7vbpld3FxFRqVG4x7GqS0ZcsPwc0V6ba5LpubwkvxY06Y5Se12k4Pcyy1LyWa4/sVTzPxkrGWW63nLNlLDWe5zUf8AcT8Ez+yxQ++RZdvU6ffVbG6VSk09sSi+JI+v6X1WHUtaptxlBrVHhr/KPhabxJN+Dt6bdVrSvKpQlpb5TWcnPLHb0YZ6mn6PJ04081XFR/qMaVw7madCOaEdnOSaz/8AHz8+Dwul3H1lxlVmpVW92/u/I+mppKCS4RzdrJI+S+l7/i0If/Kf6JHP0O3X/T/VKz5qJpfhH/k6fpHRle9ft7Wny6cU34Tbbf5GVOtCy6Nf2ieJQqVIY/b9De+tOOv8rXymc5Z29Ei5dQj4SZwy2TS8s9Loz9C4lLS6k3T2gvL8vsdL6cMf9np9UqRp2snLlcfPsfNxp6Y7nbf3NW5uGnKLhB/d4ycksfNLuxjNQ5Mt1i2k9tzN/aZfZtvGxVLKyaYFsiho9kUUQNqOJrSzbKhScZ5ymsJdzmpvTKL9zsm1mlmCmlP7LeM5CyqSTnST0vHy4MG2mlnZdj2KU6VtSU6cVOL2cXvjfdM825xKtOcIKMW9kuxFrllvItjG/DK4xLJZ8FZQ5vHJNKbw3jOhZIxjghylGMktlJYeO5DbqpX0I4UqbXuhO/jhqFPL8vY4QDbStXdaSbSWPBkSAIBpCn6jwpQT/qeDafT7qENboylD+aPxL9CbjUwys3I5QSQVhIIJAAgkAAQBIAAAAAAAAAAAACCQAAAAAAAAAAASAA2jQlJZM505Q5Q2141UEEhk7gAAAAAAAglEEgQiQAAAAAAAAABvbXMqE008GADUurt9HbdS1pKTPQpVYVOcM+Spza4Z2UbydNrc4ZYfj6fB8iTrJ9TG1pVVwhLpNOS4PLtOqbrLPaoXsZpbnC+WL6OOPDyvJueix3xE8e46VKm3pTPspSU+GYVKEZ8oTms9mf8ATcM5vF8LOjKD3TMz6656XGom0jxLrpkqbeEenDlxyfI5/g8nE8wFp05QeGsFDq8NmvaQAEAABBIAAAAAQSAAAAgkgAT3AAnDxlp48kHRRvJUaFWi4xnTqLGJdn5RzhQABAlLYgsk2tkBUvTWWxGDlLHHzNoqKTS3aCtKdtracpKK8vuJW3ozzq1EuprllvH9i1XGhLXl9sdyOvjjrpjOLbyRHJpKLaz2fgzitVXR+pYxljo3XHBONTz3N1FLsUnFJ58lY0pKnKME3wzKDcZPw1g66k4zglHZ+DKnSdaEoR+3FOaXlLn9N/wI3rvpZPVJfI67XaLb5ZxQfjujqpPS0vYSLt6lnX9C4pVE8JSWfddz76k8xWD87s6NS6rQo0knObws8fj7H6V0ey6b062hG4r1L6olvKq3oXtGPj55Zyz078e9Pjb29p0OsX13OSctXo00ucRSTPClUne3U69ODwvjqRcvtKPdn7XTuOjzWHb2iT7OjD/BzXH0W+jfUVOSsaEJTWJStpem3+C2/Qk0ZZX1p+Gentqnw98eS8K84KooNpz2bXjwfoXW/wD+M3GnKr0m6nVa39KtjV+DXJ8NW6VdW7lGVKTcXh45T+R1mUrhcbO44m3/AFY+RSbqTSWhpfIvJSi8Sck/D2KvX7mmEaGl4KtaYk4l74KvHcCre3uyeEThZzyWUd1n8giqjybtqdJp+DKLzMsnp1r2wFUpNxksNpPsdCxqWrjO5hFaaiws43waYaWWsZBouqGjLX6dznXB1JyklqeUuERKnFrL/Qm2/C62wlhJEao+nKLhlvGJeCJcZLLDSK5sMA0qxxhozIAAAg1o3FW3lqpVJQ+TMyBZtqW43cayxVbkn8T3ZnwE2nnwb1FTq09cWlNcx8k9Na8u/tgQSQVzCSCQAAAdwAAAAAAAAAAAAAAAAAAAAAAADptaDqzWxjTg5ySR9D0uwxhtGc8tR34eO55aaUOnfw02jkvrHTF7H1lKglBLBx31snB7HjnLfJ9vP4mP9t8JOGiTRU7uoUfTqPY4T2y7j4GePjloABWAAAAAAAAEEgAAAAAAAAAAABMXhm8Xk58mkJErpjW6bT2Z1UL2dJrLyjkTySjNkvt6cc7jdx71t1JSwmz06V1GS5PkYtp5Wx00b2dPGXlHDPil9PpcHzbj1k+r1Rktjnr0Y1E8o82h1FSxud8LiNRcnnuFxfVw5uPlmq8i96enlpHi1reVNvY+wnBSRw3NnGaeUdePm11Xg+V/T5n/AJYPlwdt1ZSpNtLY4mmmeyWX0/PcnHlx3WQACuYAQBJBJAEggkAQSAAAAAAAAAAAKBvCpDjgwAGsuchTafG5RS2w+C7aWNK3A2eyXkjbGc7+BCWt4k9yJSUJYW7Mum/tqlsvBSLxV1Lgt6inBpLc3pWqrQSjnX8h67rVly6xTGOTGrhPfdnp07JpfGm/ZGrs6dShKLppRfEl9pPyYvLjHfH4XJlHk0qXqRypJPw2Yy10akZRbjKLymux7Vt0qhLm4lJ+I7HmX8dNZRis42LjnjldRjl+NycOMyz+1FLXNScIptcx2/Q1jnWjCDzFNco6YfFho6PP7fRfRyh/uVsb/YT/AH/sfSRk8bs8rocNNhTf82ZfqeskcMu69eNsxkSWjOcHmE3F+zKpE4Ibrvt+tXNDCqPXH3OTrFK16k3d0Uqdyl8S/n/5M2UlHYJXj1bGhcxxVpRn81weJf8AQHSTnat//CT5+TPq5QxLKM6sFKOcGpbGbJX57nDcZJxktmn2Ikso+g6z0tVk61FJVY8/1I+ccnF4e39jrLtwyx0pKU08YJSe7b3f6EuT75RV7rZlZSnh7LYvNfFkotl7I1dOpGhCq4tQqZ0vzgCZrRPbfPJMaib42M1LZEbphZlZ6byWlORkqmdnsRLhPsQ8R92TTVzt9M6qwn4Mk2joccowaw2iuY5tpp8FcbZ7AEAAAAABBIIAkAAQSCAJAAAAAAAAAAAAAAAAAAAAAAAAJSy8EHb0+2daqnjYW67axlyuo7+mdPziTWWfUWlsoJbGNjbKEVsenFKKwfP5eTdfo/h/GnHju+0pYRhcxTgzZs4ryuowe5yx9vdn1i+W6xBang8Q9Tqlxrm0jyz6WHp+V+TZeS6AAbeYAAAAAAAABBIAAAAAAAAAAAMBbAAawmbJ5OTODWEyWO2ObcFU8ljLtKnLW6Z0UL2dN7vKOchksl9tY55Y9x71vfxmludinGaPlVJxeYvDOy36hKGFJ7eThnw/cfS4P6hZ1m9evbRqJrGTxL3p7g3KKPXo3kZpbm04Rqx7HPHLLjvb18vDxfKx3Pb4+UXF4ZB6/ULDTmUUeQ04vDPbhlMpuPznPwZcOXjQAGnnAABBIAEEgAAAAAGQAGRkAME5Yy/IEYfhmsIbZZEYyaz2L87ZKIai+wUUnnGxrCKjzuVnNLIFJLfKIUHKSS3b7BT8l4U5zadNPbv4Ism7p6VDpde3SncU8Q5a7myuNVeHp0u+0UuTlh1CtJaas20lg6PrdOUNThmXZI898r7fXw/tSf8At3X/AH26Zzm4yqcSjzHOcovTuVWjBR/I831KdKLcZSlOS3S4RFvJwg5RzqeyTJcOnXHnvlJ/8ui3lKFy05KKXOe5yXiceoOLzjlfiatxuLiEZNJJfFLyTf01WnCpTqRUo/Dg1jdZOPLj58VmP1emdnGlTu8XGPTaz57HbKwaUqtrmcEsum/tL/J56qUq84Qk4UnB71Hnc9SldWtom1XlXl/LTjoX5suXlvcceKcdnhlrX79vpukx02FFPZ6FsejE+DuOs3dw0o1PSpriFLbH48nVafSO9oYVRqvHxPn8xrJLMfqvtCTx7P6RWlxhVG6M/E+PzPWjOM4KUWmnw0yM6GZuW5M56UzmhUc5t9glbuOUZyhszWPBEsYKw8m5WM5Pk+r26p1vVivhns17n1XUZaGsdzyOqUNdo13xkuN1Uym4+cxts9iNL7YIzjfsyU3J4TwdnndPTenVepX9O2g8at5y/lj3Z39bvKVarC1toqNvafBD3Oi2qLpHQpVI/DdXvwxfeMP/AN/Y8Oo0lhEjV6j07mwtlZxr2tZzejXOLXGyyl8tziqW86cVKUWoyWU+zPT6I7a6oVKF1LTj7MnLGDosZRha1LS4jFyoycGn3XYzvTXjL2+dk2lhoqt2e5/079YuJK3uKVOEd6nqv7C8+6K9R+jdz062VzGpC5tX/wCtS4XzXY15Ri415UUtJz1Us7Gik4to9nolWyoRrK8lKM54xhbYRb6STb50H0HXb+nojbW8MJreUuceDwsrwiQs0oC6a8RGr+lBFAXUlh5Sz4wRr/pQVUFtXsiNXsEQCc+wyBAAyAAAAAAAAAAAAAAAAAAAAAAAABMVqaR9N0i20xWx4lhQdWqnjbJ9hY0VTgjz8+epp9T+n/HvJl5X076MdEUaajLWkjGrdxgt2eKdv0PWLerVUIttnz3U71JNZLX/AFVJNJnzdzdSrSe+x6OLit7r5vzPlzGeMZVajqTbZQA9r87bu7AAEAAAAAAAAAQiQAAAAAAAAAAAAAAE8AgDSE8G8ZpnKWUsEsdMc9OoGMahopJkdZltJOAtySNLU6kqbymelbXyaSlyeYhwZyxmXt34ubLiu496U41Y45PFv7LS3OC2L0rmUHu9jq9aNWPk5TG8d3Hsz5MPk46y9vAawwdl3bqLco8HGemXcfF5MLhdUABXMAAAAAB3AAAAAAABrCoow0+lTbf3pLcxLAaeo3/wUb3CD5AvDGMyf4CVRPZLYyNKdPVv2KKnoWN0lD0pJZ7M58R8FHFRlmLwZs3NOnHncMtx11nSm2l8Ml+pnGTSxgp9pZfJeGeM4RNad7l5XaVJZ+LKXsXip3NRRisJcJcJFcI1g5Z9OnnEtm/JK1j/ACq2nrjT+zH7z7ltFHTpjOrlrdtLb5FKsdFZxfKSM28/MsnTlyclmVi87ejjZzfzwZfV47/E1gl1HulnZFIttP8APDN6cLWkacoPMZ6l4ZrGq1JRnB5fjfJhFNQk4y+Jb48llJ61qjJPs4saWZ2OlTTeFLfwzptb+5spZoVZRXePKf4HF6qeMrV7lZ1XTaxnHhksdZy/r6il9Io1oabiHpz/AJov4X/g9K1q6oLG65yfEKopLPC8nZ0/qdSxrLEm6Ofihz+Ri4/jp5SvuVLYhvY5aVzGpCM4tOMllNdy06uEYNOG9SqXEY9uTivFmLR0a/Vuaku0Fg5byWwhXylRaK04PjUzqsLVVriKk1GmvinJ8RiuWYVVquaku2p/ibur6dt6MOZvNRr9Ednm+1+oXf1y5c4rFNfDTi+0UcM3l7GyWPc53sypWlGp6c09Kks7xlwzr+vSneOs4xip4TjHhHAmXbygbe9RrTuK9KnSelYcZ4+9HG8T0ruNTp3QL71KmKV2/ToUm/6ufyR5XQYyueo2sYrFOjL1Kkn+x1fTCvK5vKMozUqEIaUl92XfP6GNdum+tvHo2Oq1ndOrTlGHMVLdfgcs6qa+FfidFCo3JOKTqLhNZz5XyMLl0al1L6tBwp52i3nH/Bpz+mFapKrPVN5fBnjJ6lpZUry2qw+zVhvGa4+TPOqU5UarhJfFEbNfamCCxD4CIGCfAwBAJZAAAAAAAAAAAAAAAAAAAAAAAAAAABQtCLnJJEJZZ2W1NRkmyW6jfHhc7p63S7ZU4ps9r1404nhwvI0YcnLcdTcsqLPFcMs7t+i4ufi+Nx6j2LrqagnueLc9TlNvDOCrXlUe7MXud8OKR835Hz887qNKlWVR7szAO8mnzcsrld0AAZAAAAAAAAAABCJAAAAAAAAAAAAAAAAAAgEgBklSaIAXbaFXBqppnIFJomm5nY7kSckarRrGsnyTTrM5WojJxezKqafckje/xq561hnFWhiWVwdBWUcoTpM75ztyAtOGllTby2aAAEAAAAIAkAACASAJGGzejZXFdpU6M5fJBZLfTFIM6qvTrqj/ALlCcfmjncJJ7rHzG1uNntTDLwk488BIlrJWWsWms5Kt/EUUGW07+4VpBx1fE8LyXjKLeDLTkmK0yTayiabnJXSkksvgmlV01E4vDXcxnNz9l2Rva2dSq1KU4UaXepVeEv8AJNfrV5Lv/FWacqkpPcwSbqvdYNK9WFJ1KdOoqjUmlOK2l7nGpNZNRyttu63y5yko/C15KrZaZbPyZqo203z5LxbccdvlkIvFTbzGO0e/gZWlYe2e3gr8OHhJPykTCSSw3KfsuComMsrxh9yK32ljZYJUKjmpcCUWsObz22CqwlKDynguqkcrPwP24M5PDWMFGsv3Isunq2vU7mxxGLUqec6Xun8j1n1ulXhiDcaj+7I+Wp1ZUnwmvD4Zo5QqNKK05+63/cxcduuPI+toL0qD1fae7Z53UbpU4ZW8uyPLh1C4to6HNyivuyMaly7hqb5XKJMVyzRL4Fl89v8AJSGW8sicnJ7vLZeL22R0cUyelGEuTfQ5e78I2h027qLVG3mo/wA0lpX6g1tx6XtsaQp1K9WMKNOU54woxWWz2qH0fp+nGrfX9GhS5ljfSvGfPsjap13p3S4ypdGtHJvmtV5l/f8AYzv8amH69Tp3RKtC3hRVSNHPxVJPeUmY38+gdObjcyd7WX3FLOPy2R81edXv73KrXElB/ch8KOBR8mZjfut3PGdYx6F91iV0pUra3pWlu9nClHeS933OajCnjnLKUmo1c42SLySbyvhflGvTGre3oSs7iz6Y7inVxGosyivB48pOU25btnVVv7iVJUXLFNLTp8nJjcRKlkPgkhlZOw3CJAqyCXyQAAAAnBGD1bDpErmn6lWThB8YW7M5ZzCbrtw8OfNl44TdeZpDi0s4eM4yfRfUbS1g5SpqWPvVHk8S7ufrFX4cRpR2hFLCRjDk87078/xLwY7zvf45gAdXiAQSAAAADsAABAEgFkmwqMFlBs1jS8mmlRRLXXHj/WcYKIdVrgrUnjYybyTW1uXj1F5VW+WUbyQDWnO5WgADIAAiCe4AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAACCQAAAEqTRZVWihAalsbxreS6qpnMEyaamddMkpowlHSI1Gi0pKSBdZMwQSVzAAAIJAAAACYLMgll4RvCOle4bxx3W1CsqDT9OMn7np0fpDUoL4aMfzPHckjOVTPBi4y+3eclw9PTvuu3F099MV4R5c6s6jzJlXySnuakk9OOfJcvYmA9hvyac0lovchcBPcDaJ69l9Hb28UZaFRpy4lU2z+B4ik0XnXq1XmrVqTf9U2xd/SzX2+iu+hUbCmtF9ZzqtbutUwo/KPf8TzZdOp156rnrVmn5zKX7I81JZ4RZEmN/Vuc/HovpXTYSz/rdtNePSmclajRpzxTnTqR/miZLDeMJ/NGqo0pLOhL5GpjWbnPxVUW1lU0aOwqyoqrGjmD21RkuSf4sMaKmV4nuXV1UUJQnS+GXOndfMllhLjXHOhUpNOSqQf9SI9Sov5WvkejDqU3iEZJvj4i9ScZQaqW9Fz5yljJnbXj+PNVTO0ouPujOWze+UzrbtZxyo1IP2eUY+iqk3GGHLspbZNbZ1WD2C3L1KMqf2oSjnuUxjhgMkNjDNKFvVuaip0oOUu+O3u/AFHKU8R5a4ZvbWla6n6VtSlUljL0o+j6b0C2tqeu8nGtUnjSo/ZX+T141aFupNRwqeYtxWM/gc7n+O049+3zFl9Hp11Gpc14Uact0l8Tf9kerbdEtKcFKdJzePvzyn+Xc3neKtH+BCMqecpwm4HPedTlRpVPUnCMpLEMS1v5mbcq1McY67ajRtNdWCjCnJLG2DyeqdbpVZ/CvVcdopv4V/k8296nUuoRprMKMFhRzz8zzpS1M3MfusZcn1GlxdVbmeqrNyxwuy+SMtSxhL8SCO5tz20S2J4WSurbCIbbwgjSm1pfklbsKm+UtyEzLpMutFTbfBnCemak4xl7NbG0sOJzlhl1du2NS3qrDowg/Y5KsdE2k8rsVzgN55JJoyz8p2Ikgkrmq+SCXyQAAAHb0u0V1dLX/tw3l7+x7dz1SharTnMv5Y9jxKVy7Wz009qlV5b8I45Nt5b3PPlxf3Mt5en1eL5c+LxePHP8r7rqveoVLyW/wwXETkAO+OMxmo+dycmXJlcs7ugAK5gAAAAAAQBICWTanSzyGscbVYU3I6IU0i0YpIltIxa9GOEhwjGrUwhUq42RzttvJZGc89dRDeWADTgAAIAAAAAAAAAgkAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAQSAAAAAAAQSAATwAFAAEAAAAAAABWsElyTKquEZamyBpry61E5yyVyVRbBWdpwguSCQg9y0XhNMqFyABOdy2yWWwsm17ag7itGmpKOe8nsjW8tfqtd01NTWM6o8HJreS0ZZ7iey60t3LIgjUlyzTDVNF1Jrgw9RIlVY9s/kXaWOqMsolyOeEsPOHgtKqlyma8mfGryjGf2kmWjOrSXwtTj/LP/ACYKtFctr5lvWi1tJMl1fazc9InKLm5RTpt8wlx+BWKaWp8ou5KSw1lGLg0npeF4Zi4/jcy/XVWquq4yyltwjCc9XJEZOTw1hnbQoQptSqxTfh8RMXp0natl053eZ1JKlSW7k+X8j1Kc5WdGP1dpUXtiL3fzx/crTqYTi3JQfeE917nTChBvVOPqPhvjV4zg5279ukx16dHrKjGjOrc4pSynGa+1n9mYXdeagoas1IvaWPh52S8kObTqf9vlz7S04k/J4d3eynN06Uk1w5R2S9l/kSbXK6b3N3GnFU4S1zXLxsmedKTk2292RhJFJS7I6yacbdonPOy4KllDPOxZYS2KyootkNNFnLwE33AiL3Lpbpldk8pbhSbYHRRq6Ki1bxexStiFR4Wz3IccRcnwuPdnXKwlPpMLpPLTeV7Eajhm8xM8E5CKlu0MBgIEvggZAhkEgggkgkC05ann9PBQALbs7AAIAAAAAAAAAF6UNTz2CybaUafdm6RXUoorKskZ9vTNYxeUtKMKlXOyM5VHJlSyOeXJ+DeSACuKSCe4AAAAAAAA7gAAAAAAAAAABBICAAgkAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAFyABaWOxUAKAE4KiASCCMAkAQCQUFySIrcnAAAkCASMAQyMbFiUBXBOE02nuJcYRTJGtraW1yXhTX3v0Kwlthl1sVlOIrhEqX4EZAFm3yzKcm2Wm+xXS++y9wI1ZW+5GnPG5ZJfP5k4b2y0vZAZ7xZeMZNZbaXzKy05xHMn7msI4gs7p8kXTqttMUnJLbyTN1KlZenNZxjDMcOH3nhhVmpL90Z01t6EIzpwzUmvfTsa0Lmpqzq+CL1Sb4x7nlyuZKfnwvJnUuqlWHp5SjnOEuf8k8WvLTq6j1Od3N0qOY0uPeRyJKKEY6VnuVll/I3Jpzt2OWXhceSMxRGyJTS7FRLllbIrjyXxq7lXH3Ah4XBXJZoq0BBtFKMdT57IzUe74EpZCkpOWEexO+p0FGg45iopZPF7lnJy3f6kNtbmnGFTMHmEt0YkttoJFEEFmgkEQC2CGBXBBZkAQCWQAZBLIAAAgAAAQSAAAAJbmnqYWEZgNS6S5tkZACbAO4CAIJAAAAAAAAAAAAAAABAEgAACCQAQCAIAAAAAAAAAAAAAAAAAAAAAAAAAYAAnDGAIBbBGAIBJOCioJADGwJXAaAgE4GAAJIwAIwTgkCASRgCYokRWCQBBbAwBAJAEMJNvGC2AsxeU8AUksTxkrLHYmcnKTb5IwRUZNIz8meC2Co0ymSimrCwiMtgaJ4ZEpRby1llcY55CWpgXi0Rq1TS7e4w3suEMZXhruQWjBRntjfybfC4rb8CmPhTzl+xDk0/2QaWrSa+zjBnFYWe7J3b347IzlLLYTaJSzsvzCWle5KWFkrnJReDyn5KtsmHJDTy9giCSCUBeK2DJjwRLkCjGMbssklyQ9wIc/Y1trdV225JIxaITa3WwVetRlRnpkvxI+4iXXlKOmb1L37FAie5ZI66PTXc28alvWhOf3qecSRhOlOlJxqRcZLs0NmrGTQW5aSCQEFWXZVoCpBbBAEAACASAIBJAEAsGBUAEAAAAAAAGAAGBgAAAAAAAAAAAAAAAAAAAAAAAAAECUBAAAAnAwBAJwMAQCcACASCiASCCASMFEAkEEAkYAkAFE5IzkYAVBIGAgASBAJAEAkAAAAAABInAROQIwEi2Q8oAlktjBESwFcAtjI9N6c9gSW+lcE4ICAsRP7LLbYKz+yBm4YSfkhIsxyBUlIthLnkjOOAJSRLwirkWi4PZxfzCybV5ZdLPBTK1bPbyaxawEVWxZZcudiHjOxpFYQELK+XgnYNlGBEiuEtySrAh5l7EqK+YYi34AutiJTUZeRn2IeG+MgPUg+UNKe6/UjdbpJEPPfcC62IyI8E4AgYJwAKtEYNMFQM2gXaIALMXlPD8o1dzUksTk5LxLczZCWWBdyTCIaQQB8kMl/IgCMkMkYYFSCzIAgFiAIBIAhBkjsBUEgCASMAQQWwQBGScjBIEDIwABBOBgCCRgEEYBIwBAGCcAQBgkCAAAAwAAAAAAASiCUAwST+AKKgkgAACACSAJ2GwGCgBgYAAYJw8ZAgAAASkWUE/vYApgtpNY2+t4U1kv9TqeSbXTDTtyNCfc3+p1CPqlQbNMdC8jSvJv9TqMn6lVGzTmaQOn6jV9g7Gr7Daac6jnuQ44Oj6nV8G9t0+UqiU5Rh7sWyNY43K6jgwSo5eD2J9JrSfwypSivbByVun1IVGoL4fmZmcvp1z+PyYTeUZQtFJZdSPyMpUcPCeTf6lV8pE/UanlGtuOq5/Rl3aX4keklzJHT9Sq+xP1WouVF/gNmq5NHiSLKm5PCa/M7PQkvu0/wAUTGhN/dp/hHI2ac0rOrFZUdS/peTJprZnrQUaa+Oo4/Kl/wAnPKjZ5bdWeP8A4jZquahSVSSTnGK8s6J2tGEW1cxk/CRZULNtYrtL3izVWto+LiP4jZpw+mn95L5kaMr7f4Hq07CjNJxqKS9max6XTxnTKWPA2arx4W0Z/wDqY/8A6m07XRHGcv3jg9iHTVGPwaopi4tKlVJy0ya23JtqSx4lOynN4/bBhUpODafKPYdlXjl0lTi37nM+kXMnvKH/ANi7Z081QzwyHGUX4PUh0a6eylFfIv8A6HcvOZwz7jcNV5WlYy+Rs+x6j6HcL70CP9FrrmcBuGq8tx9i0XHGMbHpLo9b/wByCLR6LPHxVl+ERuLjbHkSj8Txx7EpHsx6PBSWuvLHtA6l0uzi3hTl82NxNWvAXyNI8HtfUqMXmEfzwYVLaCbWmLY2aeXJFGj0J2k0tkjGVrXXH7lTTja3JUTd2tbnQ2QoSp4dSnLGeAaY43wluMNex1unRck9L+SZFZKpCMaVNJpmfJ2y4fGW2uKSZC53Oj6tVbxp/Un6nXzjR+qNbcdMNsB9kdH1G4fFJ/mif9Pum8ei/wAwaYaEuJfmVbw8HV/p14v/AEZfmR/pt1/7D/Fkavf05sjJ307e8pxwreL92isrC6qScnTgn4i8DZcZrpxZG52Lply/uL8zSPR7lvfSvxG4zqvPwIx7vg9J9IrLmUEWXRp/+7D8mLYsn8PMlBNbPAwscnsR6XUUdKqQXl6clX0iT5rL8Imdxv8A5HlqEWsuol7YJ9KEd/Wj+TPXj0huKi6zSXgiXRo5y6kpfMvlGfGvGlhfeT+RCWcnsPpFGOMz58h9Iim8TQ8oeNeTGlKSyovB0SsJtJwX5tHW+mpPaqk/2FSxnJLVcvYbNPPdjWXLj+ZSVrOOMuP5nVKzgnvXyYyt6cf/AFX+Rds6YSouK3lH8GRoz95fiaulSyv4m3yDpUe1V/8A1KMdKzjUhpW/xE4UZprdEN5b2AjHuMLyABGPcEt+xGpfygCCc+yGfYCATq9l+Q1MCMDAyxkAAAAAAAAAQSAAAAgEgCASAIAAAEkAQCQACBKIP//Z" /> These are GGUF quantized versions of [Midnight Rose 70B v1.0](https://huggingface.co/sophosympatheia/Midnight-Rose-70B-v1.0). The importance matrix was trained for 100K tokens (200 batches of 512 tokens) using `wiki.train.raw`. The IQ2_XXS and IQ2_XS versions are compatible with llama.cpp, version `147b17a` or later. The IQ3_XXS requires version `f4d7e54` or later. Some model files above 50GB are split into smaller files. To concatenate them, use the `cat` command (on Windows, use PowerShell): `cat foo-Q6_K.gguf.* > foo-Q6_K.gguf`
adalib/evaluate-data-codeparrot-small-prefix
adalib
2024-01-30T15:38:08Z
1
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:codeparrot/codeparrot-small", "base_model:adapter:codeparrot/codeparrot-small", "region:us" ]
null
2024-01-30T13:53:24Z
--- library_name: peft base_model: codeparrot/codeparrot-small --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.7.1
shanhy/xlmroberta_clir_seed12_cross_translation_augmentation_val_kin_0.550
shanhy
2024-01-30T15:33:50Z
92
0
transformers
[ "transformers", "safetensors", "roberta", "text-classification", "generated_from_trainer", "base_model:FacebookAI/xlm-roberta-base", "base_model:finetune:FacebookAI/xlm-roberta-base", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-01-30T15:33:02Z
--- license: mit base_model: xlm-roberta-base tags: - generated_from_trainer model-index: - name: xlmroberta_clir_seed12_cross_translation_augmentation_val_kin results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlmroberta_clir_seed12_cross_translation_augmentation_val_kin This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.0644 - Spearman Corr: 0.5146 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 128 - seed: 12 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Spearman Corr | |:-------------:|:-----:|:----:|:---------------:|:-------------:| | No log | 0.48 | 200 | 0.0437 | 0.4319 | | No log | 0.97 | 400 | 0.0711 | 0.5352 | | No log | 1.45 | 600 | 0.0480 | 0.5505 | | No log | 1.94 | 800 | 0.0356 | 0.5106 | | 0.0371 | 2.42 | 1000 | 0.0444 | 0.5215 | | 0.0371 | 2.91 | 1200 | 0.0648 | 0.5260 | | 0.0371 | 3.39 | 1400 | 0.0694 | 0.5078 | | 0.0371 | 3.88 | 1600 | 0.0670 | 0.5449 | | 0.0225 | 4.36 | 1800 | 0.0580 | 0.5258 | | 0.0225 | 4.85 | 2000 | 0.0824 | 0.4986 | | 0.0225 | 5.33 | 2200 | 0.0644 | 0.5146 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.1.0+cu121 - Datasets 2.16.1 - Tokenizers 0.15.1
checkiejan/phi1_5-marking-test-checkpoint297
checkiejan
2024-01-30T15:24:22Z
0
0
transformers
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2024-01-30T15:23:34Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
LexFerrinson/FirstModel
LexFerrinson
2024-01-30T15:16:41Z
46
0
transformers
[ "transformers", "tf", "distilbert", "token-classification", "generated_from_keras_callback", "base_model:distilbert/distilbert-base-uncased", "base_model:finetune:distilbert/distilbert-base-uncased", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2024-01-30T14:06:02Z
--- license: apache-2.0 base_model: distilbert-base-uncased tags: - generated_from_keras_callback model-index: - name: LexFerrinson/FirstModel results: [] --- <!-- This model card has been generated automatically according to the information Keras had access to. You should probably proofread and complete it, then remove this comment. --> # LexFerrinson/FirstModel This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset. It achieves the following results on the evaluation set: - Train Loss: 0.1082 - Validation Loss: 0.2602 - Train Precision: 0.6351 - Train Recall: 0.4246 - Train F1: 0.5090 - Train Accuracy: 0.9461 - Epoch: 2 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - optimizer: {'name': 'AdamWeightDecay', 'learning_rate': {'module': 'keras.optimizers.schedules', 'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 2e-05, 'decay_steps': 636, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered_name': None}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight_decay_rate': 0.01} - training_precision: float32 ### Training results | Train Loss | Validation Loss | Train Precision | Train Recall | Train F1 | Train Accuracy | Epoch | |:----------:|:---------------:|:---------------:|:------------:|:--------:|:--------------:|:-----:| | 0.1094 | 0.2602 | 0.6351 | 0.4246 | 0.5090 | 0.9461 | 0 | | 0.1080 | 0.2602 | 0.6351 | 0.4246 | 0.5090 | 0.9461 | 1 | | 0.1082 | 0.2602 | 0.6351 | 0.4246 | 0.5090 | 0.9461 | 2 | ### Framework versions - Transformers 4.35.2 - TensorFlow 2.15.0 - Datasets 2.16.1 - Tokenizers 0.15.1
Tiabet/Tiabet-llama2-finetuned-epoch10
Tiabet
2024-01-30T15:10:12Z
1
0
peft
[ "peft", "tensorboard", "safetensors", "generated_from_trainer", "base_model:meta-llama/Llama-2-7b-chat-hf", "base_model:adapter:meta-llama/Llama-2-7b-chat-hf", "region:us" ]
null
2024-01-30T03:07:53Z
--- library_name: peft tags: - generated_from_trainer base_model: meta-llama/Llama-2-7b-chat-hf model-index: - name: Tiabet-llama2-finetuned-epoch10 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # Tiabet-llama2-finetuned-epoch10 This model is a fine-tuned version of [meta-llama/Llama-2-7b-chat-hf](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf) on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure The following `bitsandbytes` quantization config was used during training: - quant_method: bitsandbytes - load_in_8bit: False - load_in_4bit: True - llm_int8_threshold: 6.0 - llm_int8_skip_modules: None - llm_int8_enable_fp32_cpu_offload: False - llm_int8_has_fp16_weight: False - bnb_4bit_quant_type: nf4 - bnb_4bit_use_double_quant: False - bnb_4bit_compute_dtype: float16 ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 4 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: constant - lr_scheduler_warmup_ratio: 0.03 - num_epochs: 10 ### Training results ### Framework versions - PEFT 0.4.0 - Transformers 4.37.2 - Pytorch 2.1.0+cu121 - Datasets 2.16.1 - Tokenizers 0.15.1
minchyeom/MemGPT-DPO-MoE-2
minchyeom
2024-01-30T15:09:36Z
5
0
transformers
[ "transformers", "safetensors", "mixtral", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-01-30T15:06:06Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
thanhnew2001/starcoder-7b-taipy26
thanhnew2001
2024-01-30T15:07:21Z
4
0
transformers
[ "transformers", "safetensors", "gpt_bigcode", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-01-30T14:01:12Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
erens/YourNameStyleTEST
erens
2024-01-30T14:54:40Z
4
1
diffusers
[ "diffusers", "safetensors", "text-to-image", "stable-diffusion", "license:creativeml-openrail-m", "autotrain_compatible", "endpoints_compatible", "diffusers:StableDiffusionPipeline", "region:us" ]
text-to-image
2024-01-28T14:29:12Z
--- license: creativeml-openrail-m tags: - text-to-image - stable-diffusion --- ### yourname Dreambooth model trained by erens with [TheLastBen's fast-DreamBooth](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast-DreamBooth.ipynb) notebook Test the concept via A1111 Colab [fast-Colab-A1111](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast_stable_diffusion_AUTOMATIC1111.ipynb) Poorly trained from two years ago.
Harshit4199/ppo-Huggy
Harshit4199
2024-01-30T14:48:58Z
0
0
ml-agents
[ "ml-agents", "tensorboard", "onnx", "Huggy", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Huggy", "region:us" ]
reinforcement-learning
2024-01-30T14:48:54Z
--- library_name: ml-agents tags: - Huggy - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Huggy --- # **ppo** Agent playing **Huggy** This is a trained model of a **ppo** agent playing **Huggy** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: Harshit4199/ppo-Huggy 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
xutongda/adm_lsun_cat_256x256
xutongda
2024-01-30T14:22:46Z
10
0
diffusers
[ "diffusers", "license:mit", "diffusers:DDPMPipeline", "region:us" ]
null
2024-01-28T11:43:17Z
--- license: mit --- Converted into diffuser models from https://github.com/openai/guided-diffusion Should be used with this PR: https://github.com/huggingface/diffusers/pull/6730
asus4/nanosam-ort
asus4
2024-01-30T14:20:42Z
0
0
null
[ "license:apache-2.0", "region:us" ]
null
2024-01-30T14:00:50Z
--- license: apache-2.0 --- # NanoSAM ORT This repository contains ORT models converted from Nvidia's [NanoSAM](https://github.com/NVIDIA-AI-IOT/nanosam) [Original Lisence Apache-2.0](https://github.com/NVIDIA-AI-IOT/nanosam?tab=Apache-2.0-1-ov-file)
pepoo20/bettermodel_gpt2_wiki_kaggle
pepoo20
2024-01-30T14:15:36Z
199
0
transformers
[ "transformers", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-01-30T14:15:20Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
thonypythony/cover-and-wm-new-architecture-cascade-training
thonypythony
2024-01-30T14:14:40Z
0
0
null
[ "region:us" ]
null
2024-01-30T14:07:56Z
![](https://huggingface.co/thonypythony/cover-and-wm-new-architecture-cascade-training/resolve/main/0step.JPG) ## [Dataset for test](https://www.kaggle.com/datasets/thonypythony/dataset-for-a-small-test-cascade-training)
NobodyExistsOnTheInternet/code-llama-70b-python-instruct
NobodyExistsOnTheInternet
2024-01-30T14:12:20Z
54
1
transformers
[ "transformers", "safetensors", "llama", "text-generation", "mergekit", "merge", "arxiv:2311.03099", "arxiv:2306.01708", "base_model:codellama/CodeLlama-70b-Instruct-hf", "base_model:merge:codellama/CodeLlama-70b-Instruct-hf", "base_model:codellama/CodeLlama-70b-Python-hf", "base_model:merge:codellama/CodeLlama-70b-Python-hf", "base_model:meta-llama/Llama-2-70b-hf", "base_model:merge:meta-llama/Llama-2-70b-hf", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-01-30T08:23:24Z
--- base_model: - meta-llama/Llama-2-70b-hf - codellama/CodeLlama-70b-Python-hf - codellama/CodeLlama-70b-Instruct-hf tags: - mergekit - merge license: mit --- # Codellama-python-instruct This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using [meta-llama/Llama-2-70b-hf](https://huggingface.co/meta-llama/Llama-2-70b-hf) as a base. ### Models Merged The following models were included in the merge: * [codellama/CodeLlama-70b-Python-hf](https://huggingface.co/codellama/CodeLlama-70b-Python-hf) * [codellama/CodeLlama-70b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-70b-Instruct-hf) ### Configuration The following YAML configuration was used to produce this model: ```yaml models: - model: codellama/CodeLlama-70b-Python-hf parameters: density: 0.5 weight: 0.5 - model: codellama/CodeLlama-70b-Instruct-hf parameters: density: 0.5 weight: 1.0 merge_method: dare_ties base_model: meta-llama/Llama-2-70b-hf parameters: # You can uncomment and set these parameters as needed # normalize: false # int8_mask: true dtype: float16 ```
NobodyExistsOnTheInternet/clown-SUV-4x70b
NobodyExistsOnTheInternet
2024-01-30T14:11:18Z
51
4
transformers
[ "transformers", "safetensors", "mixtral", "text-generation", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-01-30T11:58:39Z
--- license: mit --- ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64b24479cb28be619964952c/AGdezCkqSp4UQBZ0OU7w6.png) The smaller brother to the clown truck. 4 clowns in an SUV. Untrained. Models used: WizardLM/WizardMath-70B-V1.0 Sao10K/Euryale-Inverted-L2-70B NobodyExistsOnTheInternet/code-llama-70b-python-instruct Technoculture/Medmerge-wizard-70b Full config can be found within the files.
kavg/LiLT-RE-EN
kavg
2024-01-30T14:09:50Z
19
0
transformers
[ "transformers", "safetensors", "lilt", "generated_from_trainer", "dataset:funsd_re", "base_model:nielsr/lilt-xlm-roberta-base", "base_model:finetune:nielsr/lilt-xlm-roberta-base", "license:mit", "endpoints_compatible", "region:us" ]
null
2024-01-30T08:39:07Z
--- license: mit base_model: nielsr/lilt-xlm-roberta-base tags: - generated_from_trainer datasets: - funsd_re metrics: - precision - recall - f1 model-index: - name: checkpoints results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # checkpoints This model is a fine-tuned version of [nielsr/lilt-xlm-roberta-base](https://huggingface.co/nielsr/lilt-xlm-roberta-base) on the funsd_re dataset. It achieves the following results on the evaluation set: - Precision: 0.3264 - Recall: 0.4864 - F1: 0.3907 - Loss: 0.4377 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 8 - eval_batch_size: 2 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - training_steps: 5000 ### Training results | Training Loss | Epoch | Step | F1 | Validation Loss | Precision | Recall | |:-------------:|:------:|:----:|:------:|:---------------:|:---------:|:------:| | 0.1604 | 26.32 | 500 | 0 | 0.1513 | 0 | 0 | | 0.1012 | 52.63 | 1000 | 0.0098 | 0.0786 | 0.5 | 0.0049 | | 0.0994 | 78.95 | 1500 | 0.2518 | 0.1847 | 0.3729 | 0.1901 | | 0.0694 | 105.26 | 2000 | 0.3499 | 0.1926 | 0.3667 | 0.3346 | | 0.0771 | 131.58 | 2500 | 0.3856 | 0.3295 | 0.3450 | 0.4370 | | 0.0565 | 157.89 | 3000 | 0.3865 | 0.4137 | 0.3293 | 0.4679 | | 0.0411 | 184.21 | 3500 | 0.3808 | 0.3624 | 0.3252 | 0.4593 | | 0.0463 | 210.53 | 4000 | 0.3832 | 0.5089 | 0.3221 | 0.4728 | | 0.0414 | 236.84 | 4500 | 0.3911 | 0.6137 | 0.3305 | 0.4790 | | 0.036 | 263.16 | 5000 | 0.3910 | 0.4428 | 0.3275 | 0.4852 | ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.16.1 - Tokenizers 0.15.1
tanatapanun/fine-tuned-BioBARTv2-20-epochs-1024-input-384-output
tanatapanun
2024-01-30T14:01:25Z
90
0
transformers
[ "transformers", "tensorboard", "safetensors", "bart", "text2text-generation", "generated_from_trainer", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2024-01-30T13:25:08Z
--- base_model: checkpoint_global_step_200000 tags: - generated_from_trainer metrics: - rouge model-index: - name: fine-tuned-BioBARTv2-20-epochs-1024-input-384-output results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # fine-tuned-BioBARTv2-20-epochs-1024-input-384-output This model is a fine-tuned version of [checkpoint_global_step_200000](https://huggingface.co/checkpoint_global_step_200000) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.6568 - Rouge1: 0.147 - Rouge2: 0.0254 - Rougel: 0.1214 - Rougelsum: 0.1199 - Gen Len: 34.9 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 20 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:| | No log | 1.0 | 151 | 6.2777 | 0.0633 | 0.0008 | 0.0582 | 0.0575 | 38.87 | | No log | 2.0 | 302 | 0.7986 | 0.0985 | 0.0246 | 0.0819 | 0.0816 | 46.96 | | No log | 3.0 | 453 | 0.6868 | 0.02 | 0.0 | 0.02 | 0.02 | 8.0 | | 3.9712 | 4.0 | 604 | 0.6495 | 0.1104 | 0.0292 | 0.0842 | 0.0839 | 28.6 | | 3.9712 | 5.0 | 755 | 0.6290 | 0.1099 | 0.0234 | 0.0796 | 0.0809 | 34.02 | | 3.9712 | 6.0 | 906 | 0.6196 | 0.1007 | 0.0201 | 0.0754 | 0.0758 | 37.81 | | 0.567 | 7.0 | 1057 | 0.6151 | 0.1086 | 0.0269 | 0.0831 | 0.0826 | 34.77 | | 0.567 | 8.0 | 1208 | 0.6124 | 0.0928 | 0.0233 | 0.077 | 0.0766 | 34.28 | | 0.567 | 9.0 | 1359 | 0.6184 | 0.0742 | 0.0243 | 0.0645 | 0.0639 | 17.83 | | 0.4091 | 10.0 | 1510 | 0.6171 | 0.1312 | 0.0312 | 0.0978 | 0.0974 | 39.69 | | 0.4091 | 11.0 | 1661 | 0.6225 | 0.1414 | 0.0258 | 0.1164 | 0.1157 | 34.01 | | 0.4091 | 12.0 | 1812 | 0.6279 | 0.125 | 0.0223 | 0.1046 | 0.1041 | 30.33 | | 0.4091 | 13.0 | 1963 | 0.6368 | 0.0677 | 0.0169 | 0.0598 | 0.0598 | 17.54 | | 0.3078 | 14.0 | 2114 | 0.6404 | 0.162 | 0.0267 | 0.1302 | 0.1279 | 39.44 | | 0.3078 | 15.0 | 2265 | 0.6417 | 0.1195 | 0.0237 | 0.0989 | 0.0989 | 33.09 | | 0.3078 | 16.0 | 2416 | 0.6491 | 0.1455 | 0.0233 | 0.1268 | 0.1262 | 26.8 | | 0.2417 | 17.0 | 2567 | 0.6532 | 0.1434 | 0.0276 | 0.1154 | 0.1152 | 40.73 | | 0.2417 | 18.0 | 2718 | 0.6527 | 0.1583 | 0.0251 | 0.1308 | 0.1295 | 36.03 | | 0.2417 | 19.0 | 2869 | 0.6565 | 0.1465 | 0.024 | 0.1213 | 0.1196 | 39.34 | | 0.2118 | 20.0 | 3020 | 0.6568 | 0.147 | 0.0254 | 0.1214 | 0.1199 | 34.9 | ### Framework versions - Transformers 4.36.2 - Pytorch 1.12.1+cu113 - Datasets 2.16.1 - Tokenizers 0.15.1
SebaHagedorn/mistral-7b-pnb-e2e
SebaHagedorn
2024-01-30T13:59:02Z
0
0
transformers
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2024-01-29T20:51:11Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
rollerhafeezh-amikom/xlm-roberta-base-fire-classification-silvanus
rollerhafeezh-amikom
2024-01-30T13:53:25Z
91
0
transformers
[ "transformers", "tensorboard", "safetensors", "xlm-roberta", "text-classification", "id", "en", "es", "it", "sk", "base_model:FacebookAI/xlm-roberta-base", "base_model:finetune:FacebookAI/xlm-roberta-base", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-01-29T23:19:28Z
--- license: mit base_model: xlm-roberta-base metrics: - accuracy model-index: - name: xlm-roberta-base-fire-classification-silvanus results: [] widget: - text: >- Kebakaran hutan dan lahan terus terjadi dan semakin meluas di Kota Palangkaraya, Kalimantan Tengah (Kalteng) pada hari Rabu, 15 Nopember 2023 20.00 WIB. Bahkan kobaran api mulai membakar pondok warga dan mendekati permukiman. BZK #RCTINews #SeputariNews #News #Karhutla #KebakaranHutan #HutanKalimantan #SILVANUS_Italian_Pilot_Testing example_title: Indonesia - text: >- Wildfire rages for a second day in Evia destroying a Natura 2000 protected pine forest. - 5:51 PM Aug 14, 2019 example_title: English - text: >- 3 nov 2023 21:57 - Incendio forestal obliga a la evacuación de hasta 850 personas cerca del pueblo de Montichelvo en Valencia. example_title: Spanish - text: >- Incendi boschivi nell'est del Paese: 2 morti e oltre 50 case distrutte nello stato del Queensland. example_title: Italian - text: >- Lesné požiare na Sicílii si vyžiadali dva ľudské životy a evakuáciu hotela http://dlvr.it/SwW3sC - 23. septembra 2023 20:57 example_title: Slovak language: - id - en - es - it - sk --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlm-roberta-base-fire-classification-silvanus This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on the Twitter (X) dataset based on the "forest fire" keyword. It achieves the following results on the evaluation set: - Loss: 0.5255 - Accuracy: 0.8884 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 3e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | No log | 1.0 | 233 | 0.5431 | 0.8670 | | No log | 2.0 | 466 | 0.5125 | 0.8670 | | 0.4162 | 3.0 | 699 | 0.5255 | 0.8884 | ### Framework versions - Transformers 4.35.0 - Pytorch 2.1.0+cu118 - Datasets 2.14.6 - Tokenizers 0.14.1
LoneStriker/CodeLlama-70b-Instruct-hf-3.5bpw-h6-exl2
LoneStriker
2024-01-30T13:29:20Z
4
0
transformers
[ "transformers", "pytorch", "safetensors", "llama", "text-generation", "llama-2", "code", "arxiv:2308.12950", "license:llama2", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-01-29T21:43:32Z
--- language: - code pipeline_tag: text-generation tags: - llama-2 license: llama2 --- # **Code Llama** Code Llama is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the repository for the 70B instruct-tuned version in the Hugging Face Transformers format. This model is designed for general code synthesis and understanding. Links to other models can be found in the index at the bottom. | | Base Model | Python | Instruct | | --- | ----------------------------------------------------------------------------- | ------------------------------------------------------------------------------------------- | ----------------------------------------------------------------------------------------------- | | 7B | [codellama/CodeLlama-7b-hf](https://huggingface.co/codellama/CodeLlama-7b-hf) | [codellama/CodeLlama-7b-Python-hf](https://huggingface.co/codellama/CodeLlama-7b-Python-hf) | [codellama/CodeLlama-7b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-7b-Instruct-hf) | | 13B | [codellama/CodeLlama-13b-hf](https://huggingface.co/codellama/CodeLlama-13b-hf) | [codellama/CodeLlama-13b-Python-hf](https://huggingface.co/codellama/CodeLlama-13b-Python-hf) | [codellama/CodeLlama-13b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-13b-Instruct-hf) | | 34B | [codellama/CodeLlama-34b-hf](https://huggingface.co/codellama/CodeLlama-34b-hf) | [codellama/CodeLlama-34b-Python-hf](https://huggingface.co/codellama/CodeLlama-34b-Python-hf) | [codellama/CodeLlama-34b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-34b-Instruct-hf) | | 70B | [codellama/CodeLlama-70b-hf](https://huggingface.co/codellama/CodeLlama-70b-hf) | [codellama/CodeLlama-70b-Python-hf](https://huggingface.co/codellama/CodeLlama-70b-Python-hf) | [codellama/CodeLlama-70b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-70b-Instruct-hf) | ## Model Use Install `transformers` ```bash pip install transformers accelerate ``` **Warning:** The 70B Instruct model has a different prompt template than the smaller versions. We'll update this repo soon. Model capabilities: - [x] Code completion. - [ ] Infilling. - [x] Instructions / chat. - [ ] Python specialist. ## Model Details *Note: Use of this model is governed by the Meta license. Meta developed and publicly released the Code Llama family of large language models (LLMs). **Model Developers** Meta **Variations** Code Llama comes in four model sizes, and three variants: * Code Llama: base models designed for general code synthesis and understanding * Code Llama - Python: designed specifically for Python * Code Llama - Instruct: for instruction following and safer deployment All variants are available in sizes of 7B, 13B, 34B, and 70B parameters. **This repository contains the Instruct version of the 70B parameters model.** **Input** Models input text only. **Output** Models generate text only. **Model Architecture** Code Llama is an auto-regressive language model that uses an optimized transformer architecture. It was fine-tuned with up to 16k tokens. This variant **does not** support long context of up to 100k tokens. **Model Dates** Code Llama and its variants have been trained between January 2023 and January 2024. **Status** This is a static model trained on an offline dataset. Future versions of Code Llama - Instruct will be released as we improve model safety with community feedback. **License** A custom commercial license is available at: [https://ai.meta.com/resources/models-and-libraries/llama-downloads/](https://ai.meta.com/resources/models-and-libraries/llama-downloads/) **Research Paper** More information can be found in the paper "[Code Llama: Open Foundation Models for Code](https://ai.meta.com/research/publications/code-llama-open-foundation-models-for-code/)" or its [arXiv page](https://arxiv.org/abs/2308.12950). ## Intended Use **Intended Use Cases** Code Llama and its variants are intended for commercial and research use in English and relevant programming languages. The base model Code Llama can be adapted for a variety of code synthesis and understanding tasks, Code Llama - Python is designed specifically to handle the Python programming language, and Code Llama - Instruct is intended to be safer to use for code assistant and generation applications. **Out-of-Scope Uses** Use in any manner that violates applicable laws or regulations (including trade compliance laws). Use in languages other than English. Use in any other way that is prohibited by the Acceptable Use Policy and Licensing Agreement for Code Llama and its variants. ## Hardware and Software **Training Factors** We used custom training libraries. The training and fine-tuning of the released models have been performed Meta’s Research Super Cluster.\\**Carbon Footprint** In aggregate, training all 12 Code Llama models required 1400K GPU hours of computation on hardware of type A100-80GB (TDP of 350-400W). Estimated total emissions were 228.55 tCO2eq, 100% of which were offset by Meta’s sustainability program. ## Evaluation Results See evaluations for the main models and detailed ablations in Section 3 and safety evaluations in Section 4 of the research paper. ## Ethical Considerations and Limitations Code Llama and its variants are a new technology that carries risks with use. Testing conducted to date has been in English, and has not covered, nor could it cover all scenarios. For these reasons, as with all LLMs, Code Llama’s potential outputs cannot be predicted in advance, and the model may in some instances produce inaccurate or objectionable responses to user prompts. Therefore, before deploying any applications of Code Llama, developers should perform safety testing and tuning tailored to their specific applications of the model. Please see the Responsible Use Guide available available at [https://ai.meta.com/llama/responsible-use-guide](https://ai.meta.com/llama/responsible-use-guide).
LoneStriker/CodeLlama-70b-Instruct-hf-6.0bpw-h6-exl2
LoneStriker
2024-01-30T13:27:10Z
7
0
transformers
[ "transformers", "pytorch", "safetensors", "llama", "text-generation", "llama-2", "code", "arxiv:2308.12950", "license:llama2", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-01-30T03:26:00Z
--- language: - code pipeline_tag: text-generation tags: - llama-2 license: llama2 --- # **Code Llama** Code Llama is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the repository for the 70B instruct-tuned version in the Hugging Face Transformers format. This model is designed for general code synthesis and understanding. Links to other models can be found in the index at the bottom. | | Base Model | Python | Instruct | | --- | ----------------------------------------------------------------------------- | ------------------------------------------------------------------------------------------- | ----------------------------------------------------------------------------------------------- | | 7B | [codellama/CodeLlama-7b-hf](https://huggingface.co/codellama/CodeLlama-7b-hf) | [codellama/CodeLlama-7b-Python-hf](https://huggingface.co/codellama/CodeLlama-7b-Python-hf) | [codellama/CodeLlama-7b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-7b-Instruct-hf) | | 13B | [codellama/CodeLlama-13b-hf](https://huggingface.co/codellama/CodeLlama-13b-hf) | [codellama/CodeLlama-13b-Python-hf](https://huggingface.co/codellama/CodeLlama-13b-Python-hf) | [codellama/CodeLlama-13b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-13b-Instruct-hf) | | 34B | [codellama/CodeLlama-34b-hf](https://huggingface.co/codellama/CodeLlama-34b-hf) | [codellama/CodeLlama-34b-Python-hf](https://huggingface.co/codellama/CodeLlama-34b-Python-hf) | [codellama/CodeLlama-34b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-34b-Instruct-hf) | | 70B | [codellama/CodeLlama-70b-hf](https://huggingface.co/codellama/CodeLlama-70b-hf) | [codellama/CodeLlama-70b-Python-hf](https://huggingface.co/codellama/CodeLlama-70b-Python-hf) | [codellama/CodeLlama-70b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-70b-Instruct-hf) | ## Model Use Install `transformers` ```bash pip install transformers accelerate ``` **Warning:** The 70B Instruct model has a different prompt template than the smaller versions. We'll update this repo soon. Model capabilities: - [x] Code completion. - [ ] Infilling. - [x] Instructions / chat. - [ ] Python specialist. ## Model Details *Note: Use of this model is governed by the Meta license. Meta developed and publicly released the Code Llama family of large language models (LLMs). **Model Developers** Meta **Variations** Code Llama comes in four model sizes, and three variants: * Code Llama: base models designed for general code synthesis and understanding * Code Llama - Python: designed specifically for Python * Code Llama - Instruct: for instruction following and safer deployment All variants are available in sizes of 7B, 13B, 34B, and 70B parameters. **This repository contains the Instruct version of the 70B parameters model.** **Input** Models input text only. **Output** Models generate text only. **Model Architecture** Code Llama is an auto-regressive language model that uses an optimized transformer architecture. It was fine-tuned with up to 16k tokens. This variant **does not** support long context of up to 100k tokens. **Model Dates** Code Llama and its variants have been trained between January 2023 and January 2024. **Status** This is a static model trained on an offline dataset. Future versions of Code Llama - Instruct will be released as we improve model safety with community feedback. **License** A custom commercial license is available at: [https://ai.meta.com/resources/models-and-libraries/llama-downloads/](https://ai.meta.com/resources/models-and-libraries/llama-downloads/) **Research Paper** More information can be found in the paper "[Code Llama: Open Foundation Models for Code](https://ai.meta.com/research/publications/code-llama-open-foundation-models-for-code/)" or its [arXiv page](https://arxiv.org/abs/2308.12950). ## Intended Use **Intended Use Cases** Code Llama and its variants are intended for commercial and research use in English and relevant programming languages. The base model Code Llama can be adapted for a variety of code synthesis and understanding tasks, Code Llama - Python is designed specifically to handle the Python programming language, and Code Llama - Instruct is intended to be safer to use for code assistant and generation applications. **Out-of-Scope Uses** Use in any manner that violates applicable laws or regulations (including trade compliance laws). Use in languages other than English. Use in any other way that is prohibited by the Acceptable Use Policy and Licensing Agreement for Code Llama and its variants. ## Hardware and Software **Training Factors** We used custom training libraries. The training and fine-tuning of the released models have been performed Meta’s Research Super Cluster.\\**Carbon Footprint** In aggregate, training all 12 Code Llama models required 1400K GPU hours of computation on hardware of type A100-80GB (TDP of 350-400W). Estimated total emissions were 228.55 tCO2eq, 100% of which were offset by Meta’s sustainability program. ## Evaluation Results See evaluations for the main models and detailed ablations in Section 3 and safety evaluations in Section 4 of the research paper. ## Ethical Considerations and Limitations Code Llama and its variants are a new technology that carries risks with use. Testing conducted to date has been in English, and has not covered, nor could it cover all scenarios. For these reasons, as with all LLMs, Code Llama’s potential outputs cannot be predicted in advance, and the model may in some instances produce inaccurate or objectionable responses to user prompts. Therefore, before deploying any applications of Code Llama, developers should perform safety testing and tuning tailored to their specific applications of the model. Please see the Responsible Use Guide available available at [https://ai.meta.com/llama/responsible-use-guide](https://ai.meta.com/llama/responsible-use-guide).
LoneStriker/CodeLlama-70b-Instruct-hf-5.0bpw-h6-exl2
LoneStriker
2024-01-30T13:25:12Z
6
6
transformers
[ "transformers", "pytorch", "safetensors", "llama", "text-generation", "llama-2", "code", "arxiv:2308.12950", "license:llama2", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-01-29T22:31:39Z
--- language: - code pipeline_tag: text-generation tags: - llama-2 license: llama2 --- # **Code Llama** Code Llama is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the repository for the 70B instruct-tuned version in the Hugging Face Transformers format. This model is designed for general code synthesis and understanding. Links to other models can be found in the index at the bottom. | | Base Model | Python | Instruct | | --- | ----------------------------------------------------------------------------- | ------------------------------------------------------------------------------------------- | ----------------------------------------------------------------------------------------------- | | 7B | [codellama/CodeLlama-7b-hf](https://huggingface.co/codellama/CodeLlama-7b-hf) | [codellama/CodeLlama-7b-Python-hf](https://huggingface.co/codellama/CodeLlama-7b-Python-hf) | [codellama/CodeLlama-7b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-7b-Instruct-hf) | | 13B | [codellama/CodeLlama-13b-hf](https://huggingface.co/codellama/CodeLlama-13b-hf) | [codellama/CodeLlama-13b-Python-hf](https://huggingface.co/codellama/CodeLlama-13b-Python-hf) | [codellama/CodeLlama-13b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-13b-Instruct-hf) | | 34B | [codellama/CodeLlama-34b-hf](https://huggingface.co/codellama/CodeLlama-34b-hf) | [codellama/CodeLlama-34b-Python-hf](https://huggingface.co/codellama/CodeLlama-34b-Python-hf) | [codellama/CodeLlama-34b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-34b-Instruct-hf) | | 70B | [codellama/CodeLlama-70b-hf](https://huggingface.co/codellama/CodeLlama-70b-hf) | [codellama/CodeLlama-70b-Python-hf](https://huggingface.co/codellama/CodeLlama-70b-Python-hf) | [codellama/CodeLlama-70b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-70b-Instruct-hf) | ## Model Use Install `transformers` ```bash pip install transformers accelerate ``` **Warning:** The 70B Instruct model has a different prompt template than the smaller versions. We'll update this repo soon. Model capabilities: - [x] Code completion. - [ ] Infilling. - [x] Instructions / chat. - [ ] Python specialist. ## Model Details *Note: Use of this model is governed by the Meta license. Meta developed and publicly released the Code Llama family of large language models (LLMs). **Model Developers** Meta **Variations** Code Llama comes in four model sizes, and three variants: * Code Llama: base models designed for general code synthesis and understanding * Code Llama - Python: designed specifically for Python * Code Llama - Instruct: for instruction following and safer deployment All variants are available in sizes of 7B, 13B, 34B, and 70B parameters. **This repository contains the Instruct version of the 70B parameters model.** **Input** Models input text only. **Output** Models generate text only. **Model Architecture** Code Llama is an auto-regressive language model that uses an optimized transformer architecture. It was fine-tuned with up to 16k tokens. This variant **does not** support long context of up to 100k tokens. **Model Dates** Code Llama and its variants have been trained between January 2023 and January 2024. **Status** This is a static model trained on an offline dataset. Future versions of Code Llama - Instruct will be released as we improve model safety with community feedback. **License** A custom commercial license is available at: [https://ai.meta.com/resources/models-and-libraries/llama-downloads/](https://ai.meta.com/resources/models-and-libraries/llama-downloads/) **Research Paper** More information can be found in the paper "[Code Llama: Open Foundation Models for Code](https://ai.meta.com/research/publications/code-llama-open-foundation-models-for-code/)" or its [arXiv page](https://arxiv.org/abs/2308.12950). ## Intended Use **Intended Use Cases** Code Llama and its variants are intended for commercial and research use in English and relevant programming languages. The base model Code Llama can be adapted for a variety of code synthesis and understanding tasks, Code Llama - Python is designed specifically to handle the Python programming language, and Code Llama - Instruct is intended to be safer to use for code assistant and generation applications. **Out-of-Scope Uses** Use in any manner that violates applicable laws or regulations (including trade compliance laws). Use in languages other than English. Use in any other way that is prohibited by the Acceptable Use Policy and Licensing Agreement for Code Llama and its variants. ## Hardware and Software **Training Factors** We used custom training libraries. The training and fine-tuning of the released models have been performed Meta’s Research Super Cluster.\\**Carbon Footprint** In aggregate, training all 12 Code Llama models required 1400K GPU hours of computation on hardware of type A100-80GB (TDP of 350-400W). Estimated total emissions were 228.55 tCO2eq, 100% of which were offset by Meta’s sustainability program. ## Evaluation Results See evaluations for the main models and detailed ablations in Section 3 and safety evaluations in Section 4 of the research paper. ## Ethical Considerations and Limitations Code Llama and its variants are a new technology that carries risks with use. Testing conducted to date has been in English, and has not covered, nor could it cover all scenarios. For these reasons, as with all LLMs, Code Llama’s potential outputs cannot be predicted in advance, and the model may in some instances produce inaccurate or objectionable responses to user prompts. Therefore, before deploying any applications of Code Llama, developers should perform safety testing and tuning tailored to their specific applications of the model. Please see the Responsible Use Guide available available at [https://ai.meta.com/llama/responsible-use-guide](https://ai.meta.com/llama/responsible-use-guide).
taki0112/lora-trained-xl_mixed-media-arts_split
taki0112
2024-01-30T13:24:18Z
5
1
diffusers
[ "diffusers", "tensorboard", "stable-diffusion-xl", "stable-diffusion-xl-diffusers", "text-to-image", "lora", "template:sd-lora", "base_model:stabilityai/stable-diffusion-xl-base-1.0", "base_model:adapter:stabilityai/stable-diffusion-xl-base-1.0", "license:openrail++", "region:us" ]
text-to-image
2024-01-21T14:50:42Z
--- tags: - stable-diffusion-xl - stable-diffusion-xl-diffusers - text-to-image - diffusers - lora - template:sd-lora widget: - text: 'a shopping cart in sks style' output: url: "image_0.png" - text: 'a shopping cart in sks style' output: url: "image_1.png" - text: 'a shopping cart in sks style' output: url: "image_2.png" - text: 'a shopping cart in sks style' output: url: "image_3.png" base_model: stabilityai/stable-diffusion-xl-base-1.0 instance_prompt: a cactus in sks style license: openrail++ --- # SDXL LoRA DreamBooth - taki0112/lora-trained-xl_mixed-media-arts_split <Gallery /> ## Model description These are taki0112/lora-trained-xl_mixed-media-arts_split LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0. The weights were trained using [DreamBooth](https://dreambooth.github.io/). LoRA for the text encoder was enabled: False. Special VAE used for training: madebyollin/sdxl-vae-fp16-fix. ## Trigger words You should use a cactus in sks style to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. [Download](taki0112/lora-trained-xl_mixed-media-arts_split/tree/main) them in the Files & versions tab.
ryusangwon/bart-base-chinese-6615-chinese
ryusangwon
2024-01-30T13:24:12Z
90
0
transformers
[ "transformers", "safetensors", "bart", "text2text-generation", "generated_from_trainer", "dataset:xlsum", "base_model:fnlp/bart-base-chinese", "base_model:finetune:fnlp/bart-base-chinese", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2024-01-30T04:05:26Z
--- base_model: fnlp/bart-base-chinese tags: - generated_from_trainer datasets: - xlsum metrics: - rouge model-index: - name: bart-base-chinese-6615-chinese results: - task: name: Sequence-to-sequence Language Modeling type: text2text-generation dataset: name: xlsum type: xlsum config: chinese_traditional split: validation args: chinese_traditional metrics: - name: Rouge1 type: rouge value: 0.0774 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bart-base-chinese-6615-chinese This model is a fine-tuned version of [fnlp/bart-base-chinese](https://huggingface.co/fnlp/bart-base-chinese) on the xlsum dataset. It achieves the following results on the evaluation set: - Loss: 0.8576 - Rouge1: 0.0774 - Rouge2: 0.0179 - Rougel: 0.0772 - Rougelsum: 0.077 - Gen Len: 19.9552 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - gradient_accumulation_steps: 16 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - num_epochs: 10 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:| | 1.1216 | 0.86 | 500 | 0.9150 | 0.0523 | 0.0113 | 0.0521 | 0.052 | 19.9345 | | 1.0346 | 1.71 | 1000 | 0.8817 | 0.0585 | 0.0119 | 0.0583 | 0.0582 | 19.9535 | | 1.0063 | 2.57 | 1500 | 0.8624 | 0.0603 | 0.0112 | 0.0598 | 0.0599 | 19.9512 | | 0.9219 | 3.42 | 2000 | 0.8592 | 0.0715 | 0.0145 | 0.071 | 0.0712 | 19.9535 | | 0.8757 | 4.28 | 2500 | 0.8577 | 0.072 | 0.0153 | 0.0717 | 0.0717 | 19.9636 | | 0.8832 | 5.14 | 3000 | 0.8567 | 0.0721 | 0.0157 | 0.0717 | 0.0718 | 19.9493 | | 0.8788 | 5.99 | 3500 | 0.8498 | 0.0763 | 0.0173 | 0.0759 | 0.0759 | 19.9565 | | 0.8659 | 6.85 | 4000 | 0.8513 | 0.076 | 0.017 | 0.0756 | 0.0754 | 19.9546 | | 0.7802 | 7.71 | 4500 | 0.8563 | 0.0772 | 0.0185 | 0.077 | 0.0768 | 19.9525 | | 0.8114 | 8.56 | 5000 | 0.8562 | 0.0769 | 0.0169 | 0.0766 | 0.0764 | 19.954 | | 0.7715 | 9.42 | 5500 | 0.8576 | 0.0774 | 0.0179 | 0.0772 | 0.077 | 19.9552 | ### Framework versions - Transformers 4.36.2 - Pytorch 2.1.2+cu121 - Datasets 2.15.0 - Tokenizers 0.15.0
samot-samoe/rugpt_gazeta-sft-6000-steps-lora-v01
samot-samoe
2024-01-30T13:24:05Z
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:IlyaGusev/rugpt3medium_sum_gazeta", "base_model:adapter:IlyaGusev/rugpt3medium_sum_gazeta", "region:us" ]
null
2024-01-30T13:24:00Z
--- library_name: peft base_model: IlyaGusev/rugpt3medium_sum_gazeta --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.8.1
LoneStriker/CodeLlama-70b-Instruct-hf-4.0bpw-h6-exl2
LoneStriker
2024-01-30T13:21:58Z
6
1
transformers
[ "transformers", "pytorch", "safetensors", "llama", "text-generation", "llama-2", "code", "arxiv:2308.12950", "license:llama2", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-01-29T21:58:26Z
--- language: - code pipeline_tag: text-generation tags: - llama-2 license: llama2 --- # **Code Llama** Code Llama is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the repository for the 70B instruct-tuned version in the Hugging Face Transformers format. This model is designed for general code synthesis and understanding. Links to other models can be found in the index at the bottom. | | Base Model | Python | Instruct | | --- | ----------------------------------------------------------------------------- | ------------------------------------------------------------------------------------------- | ----------------------------------------------------------------------------------------------- | | 7B | [codellama/CodeLlama-7b-hf](https://huggingface.co/codellama/CodeLlama-7b-hf) | [codellama/CodeLlama-7b-Python-hf](https://huggingface.co/codellama/CodeLlama-7b-Python-hf) | [codellama/CodeLlama-7b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-7b-Instruct-hf) | | 13B | [codellama/CodeLlama-13b-hf](https://huggingface.co/codellama/CodeLlama-13b-hf) | [codellama/CodeLlama-13b-Python-hf](https://huggingface.co/codellama/CodeLlama-13b-Python-hf) | [codellama/CodeLlama-13b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-13b-Instruct-hf) | | 34B | [codellama/CodeLlama-34b-hf](https://huggingface.co/codellama/CodeLlama-34b-hf) | [codellama/CodeLlama-34b-Python-hf](https://huggingface.co/codellama/CodeLlama-34b-Python-hf) | [codellama/CodeLlama-34b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-34b-Instruct-hf) | | 70B | [codellama/CodeLlama-70b-hf](https://huggingface.co/codellama/CodeLlama-70b-hf) | [codellama/CodeLlama-70b-Python-hf](https://huggingface.co/codellama/CodeLlama-70b-Python-hf) | [codellama/CodeLlama-70b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-70b-Instruct-hf) | ## Model Use Install `transformers` ```bash pip install transformers accelerate ``` **Warning:** The 70B Instruct model has a different prompt template than the smaller versions. We'll update this repo soon. Model capabilities: - [x] Code completion. - [ ] Infilling. - [x] Instructions / chat. - [ ] Python specialist. ## Model Details *Note: Use of this model is governed by the Meta license. Meta developed and publicly released the Code Llama family of large language models (LLMs). **Model Developers** Meta **Variations** Code Llama comes in four model sizes, and three variants: * Code Llama: base models designed for general code synthesis and understanding * Code Llama - Python: designed specifically for Python * Code Llama - Instruct: for instruction following and safer deployment All variants are available in sizes of 7B, 13B, 34B, and 70B parameters. **This repository contains the Instruct version of the 70B parameters model.** **Input** Models input text only. **Output** Models generate text only. **Model Architecture** Code Llama is an auto-regressive language model that uses an optimized transformer architecture. It was fine-tuned with up to 16k tokens. This variant **does not** support long context of up to 100k tokens. **Model Dates** Code Llama and its variants have been trained between January 2023 and January 2024. **Status** This is a static model trained on an offline dataset. Future versions of Code Llama - Instruct will be released as we improve model safety with community feedback. **License** A custom commercial license is available at: [https://ai.meta.com/resources/models-and-libraries/llama-downloads/](https://ai.meta.com/resources/models-and-libraries/llama-downloads/) **Research Paper** More information can be found in the paper "[Code Llama: Open Foundation Models for Code](https://ai.meta.com/research/publications/code-llama-open-foundation-models-for-code/)" or its [arXiv page](https://arxiv.org/abs/2308.12950). ## Intended Use **Intended Use Cases** Code Llama and its variants are intended for commercial and research use in English and relevant programming languages. The base model Code Llama can be adapted for a variety of code synthesis and understanding tasks, Code Llama - Python is designed specifically to handle the Python programming language, and Code Llama - Instruct is intended to be safer to use for code assistant and generation applications. **Out-of-Scope Uses** Use in any manner that violates applicable laws or regulations (including trade compliance laws). Use in languages other than English. Use in any other way that is prohibited by the Acceptable Use Policy and Licensing Agreement for Code Llama and its variants. ## Hardware and Software **Training Factors** We used custom training libraries. The training and fine-tuning of the released models have been performed Meta’s Research Super Cluster.\\**Carbon Footprint** In aggregate, training all 12 Code Llama models required 1400K GPU hours of computation on hardware of type A100-80GB (TDP of 350-400W). Estimated total emissions were 228.55 tCO2eq, 100% of which were offset by Meta’s sustainability program. ## Evaluation Results See evaluations for the main models and detailed ablations in Section 3 and safety evaluations in Section 4 of the research paper. ## Ethical Considerations and Limitations Code Llama and its variants are a new technology that carries risks with use. Testing conducted to date has been in English, and has not covered, nor could it cover all scenarios. For these reasons, as with all LLMs, Code Llama’s potential outputs cannot be predicted in advance, and the model may in some instances produce inaccurate or objectionable responses to user prompts. Therefore, before deploying any applications of Code Llama, developers should perform safety testing and tuning tailored to their specific applications of the model. Please see the Responsible Use Guide available available at [https://ai.meta.com/llama/responsible-use-guide](https://ai.meta.com/llama/responsible-use-guide).
Log95/donut_demo
Log95
2024-01-30T13:21:38Z
33
0
transformers
[ "transformers", "safetensors", "vision-encoder-decoder", "image-text-to-text", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
image-text-to-text
2024-01-30T13:19:31Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
AptaArkana/indonesian_hoax_classification_indobertweet_base_uncased
AptaArkana
2024-01-30T13:21:34Z
33
0
transformers
[ "transformers", "tensorboard", "safetensors", "bert", "text-classification", "generated_from_trainer", "base_model:indolem/indobertweet-base-uncased", "base_model:finetune:indolem/indobertweet-base-uncased", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-01-26T12:08:44Z
--- license: apache-2.0 base_model: indolem/indobertweet-base-uncased tags: - generated_from_trainer metrics: - accuracy - precision - recall - f1 model-index: - name: hoaxpemilu results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # hoaxpemilu This model is a fine-tuned version of [indolem/indobertweet-base-uncased](https://huggingface.co/indolem/indobertweet-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.3001 - Accuracy: 0.9607 - Precision: 0.9585 - Recall: 0.9569 - F1: 0.9577 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 | |:-------------:|:-----:|:-----:|:---------------:|:--------:|:---------:|:------:|:------:| | 0.2232 | 1.0 | 2593 | 0.2833 | 0.9368 | 0.9724 | 0.8894 | 0.9290 | | 0.1317 | 2.0 | 5186 | 0.2334 | 0.9580 | 0.9482 | 0.9623 | 0.9552 | | 0.0515 | 3.0 | 7779 | 0.2747 | 0.9512 | 0.9663 | 0.9275 | 0.9465 | | 0.0236 | 4.0 | 10372 | 0.3100 | 0.9587 | 0.9583 | 0.9528 | 0.9555 | | 0.0089 | 5.0 | 12965 | 0.3001 | 0.9607 | 0.9585 | 0.9569 | 0.9577 | ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.16.1 - Tokenizers 0.15.1
asun17904/anliR3-bert-base-uncased-kd
asun17904
2024-01-30T13:19:07Z
0
0
pytorch
[ "pytorch", "en", "license:mit", "region:us" ]
null
2024-01-28T22:17:17Z
--- language: en license: mit library_name: pytorch --- # Knowledge Continuity Regularized Network Dataset: ANLI Round: None Trainer Hyperparameters: - `lr` = 5e-05 - `per_device_batch_size` = 32 - `gradient_accumulation_steps` = 1 - `weight_decay` = 0.0 - `seed` = 42 Regularization Hyperparameters - `numerical stability denominator constant` = 0.01 - `lambda` = 0.01 - `alpha` = 2.0 - `beta` = 1.0 Extended Logs: |eval_loss|eval_accuracy|epoch| |--|--|--| |35.436|0.413|1.0| |35.895|0.412|2.0| |34.939|0.440|3.0| |34.845|0.451|4.0| |35.466|0.432|5.0| |34.967|0.444|6.0| |34.487|0.463|7.0| |35.464|0.429|8.0| |34.825|0.458|9.0| |35.149|0.442|10.0| |35.293|0.442|11.0| |34.976|0.453|12.0| |35.004|0.450|13.0| |35.351|0.441|14.0| |35.385|0.438|15.0| |35.321|0.441|16.0| |35.273|0.441|17.0| |34.948|0.455|18.0| |35.225|0.443|19.0| **Test Accuracy: 0.453**
Augustya07/Mistral-7B-Instruct-v0.2-sft-test-push-adapters
Augustya07
2024-01-30T13:18:00Z
0
0
transformers
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2024-01-30T13:17:07Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
erens/GuiltyGearStriveStyleV3
erens
2024-01-30T13:12:58Z
0
1
diffusers
[ "diffusers", "safetensors", "text-to-image", "stable-diffusion", "license:creativeml-openrail-m", "autotrain_compatible", "endpoints_compatible", "diffusers:StableDiffusionPipeline", "region:us" ]
text-to-image
2024-01-30T12:36:32Z
--- license: creativeml-openrail-m tags: - text-to-image - stable-diffusion --- ### guiltygear Dreambooth model trained by erens with [TheLastBen's fast-DreamBooth](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast-DreamBooth.ipynb) notebook Test the concept via A1111 Colab [fast-Colab-A1111](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast_stable_diffusion_AUTOMATIC1111.ipynb) The final version Sample pictures of this concept: ![image/png](https://cdn-uploads.huggingface.co/production/uploads/635cf1664fabde0df7409c28/hagxrp47_0_hvcs5GViWm.png) 1girl, high quality, Mikasa Ackerman, shingeki no kyojin, red scarf, black eyes, portrait, solo, (detailed face, detailed eyes), looking at viewer, shiny skin, arc system works, 30 steps, 8 CFG
Conrad747/luganda-ner-v6
Conrad747
2024-01-30T13:10:30Z
13
0
transformers
[ "transformers", "pytorch", "safetensors", "xlm-roberta", "token-classification", "generated_from_trainer", "dataset:lg-ner", "base_model:FacebookAI/xlm-roberta-base", "base_model:finetune:FacebookAI/xlm-roberta-base", "license:mit", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2023-03-31T07:10:15Z
--- license: mit base_model: xlm-roberta-base tags: - generated_from_trainer datasets: - lg-ner metrics: - precision - recall - f1 - accuracy model-index: - name: luganda-ner-v6 results: - task: name: Token Classification type: token-classification dataset: name: lg-ner type: lg-ner config: lug split: test args: lug metrics: - name: Precision type: precision value: 0.8029689608636977 - name: Recall type: recall value: 0.7991940899932841 - name: F1 type: f1 value: 0.8010770784247729 - name: Accuracy type: accuracy value: 0.9467474952809641 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # luganda-ner-v6 This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on the lg-ner dataset. It achieves the following results on the evaluation set: - Loss: 0.2811 - Precision: 0.8030 - Recall: 0.7992 - F1: 0.8011 - Accuracy: 0.9467 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | No log | 1.0 | 261 | 0.5150 | 0.4947 | 0.2841 | 0.3609 | 0.8692 | | 0.6193 | 2.0 | 522 | 0.3422 | 0.7491 | 0.5393 | 0.6271 | 0.9161 | | 0.6193 | 3.0 | 783 | 0.2737 | 0.7744 | 0.6595 | 0.7124 | 0.9306 | | 0.2505 | 4.0 | 1044 | 0.3201 | 0.7343 | 0.7072 | 0.7205 | 0.9141 | | 0.2505 | 5.0 | 1305 | 0.2564 | 0.7887 | 0.7569 | 0.7724 | 0.9375 | | 0.1474 | 6.0 | 1566 | 0.2461 | 0.8173 | 0.7569 | 0.7859 | 0.9459 | | 0.1474 | 7.0 | 1827 | 0.2739 | 0.8004 | 0.7757 | 0.7879 | 0.9434 | | 0.0956 | 8.0 | 2088 | 0.2566 | 0.8100 | 0.7905 | 0.8001 | 0.9486 | | 0.0956 | 9.0 | 2349 | 0.2709 | 0.7859 | 0.7938 | 0.7898 | 0.9463 | | 0.0712 | 10.0 | 2610 | 0.2811 | 0.8030 | 0.7992 | 0.8011 | 0.9467 | ### Framework versions - Transformers 4.34.1 - Pytorch 2.1.0+cu118 - Datasets 2.14.6 - Tokenizers 0.14.1
LoneStriker/CodeLlama-70b-Instruct-hf-2.4bpw-h6-exl2
LoneStriker
2024-01-30T13:09:11Z
6
1
transformers
[ "transformers", "pytorch", "safetensors", "llama", "text-generation", "llama-2", "code", "arxiv:2308.12950", "license:llama2", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-01-30T12:59:47Z
--- language: - code pipeline_tag: text-generation tags: - llama-2 license: llama2 --- # **Code Llama** Code Llama is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the repository for the 70B instruct-tuned version in the Hugging Face Transformers format. This model is designed for general code synthesis and understanding. Links to other models can be found in the index at the bottom. | | Base Model | Python | Instruct | | --- | ----------------------------------------------------------------------------- | ------------------------------------------------------------------------------------------- | ----------------------------------------------------------------------------------------------- | | 7B | [codellama/CodeLlama-7b-hf](https://huggingface.co/codellama/CodeLlama-7b-hf) | [codellama/CodeLlama-7b-Python-hf](https://huggingface.co/codellama/CodeLlama-7b-Python-hf) | [codellama/CodeLlama-7b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-7b-Instruct-hf) | | 13B | [codellama/CodeLlama-13b-hf](https://huggingface.co/codellama/CodeLlama-13b-hf) | [codellama/CodeLlama-13b-Python-hf](https://huggingface.co/codellama/CodeLlama-13b-Python-hf) | [codellama/CodeLlama-13b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-13b-Instruct-hf) | | 34B | [codellama/CodeLlama-34b-hf](https://huggingface.co/codellama/CodeLlama-34b-hf) | [codellama/CodeLlama-34b-Python-hf](https://huggingface.co/codellama/CodeLlama-34b-Python-hf) | [codellama/CodeLlama-34b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-34b-Instruct-hf) | | 70B | [codellama/CodeLlama-70b-hf](https://huggingface.co/codellama/CodeLlama-70b-hf) | [codellama/CodeLlama-70b-Python-hf](https://huggingface.co/codellama/CodeLlama-70b-Python-hf) | [codellama/CodeLlama-70b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-70b-Instruct-hf) | ## Model Use Install `transformers` ```bash pip install transformers accelerate ``` **Warning:** The 70B Instruct model has a different prompt template than the smaller versions. We'll update this repo soon. Model capabilities: - [x] Code completion. - [ ] Infilling. - [x] Instructions / chat. - [ ] Python specialist. ## Model Details *Note: Use of this model is governed by the Meta license. Meta developed and publicly released the Code Llama family of large language models (LLMs). **Model Developers** Meta **Variations** Code Llama comes in four model sizes, and three variants: * Code Llama: base models designed for general code synthesis and understanding * Code Llama - Python: designed specifically for Python * Code Llama - Instruct: for instruction following and safer deployment All variants are available in sizes of 7B, 13B, 34B, and 70B parameters. **This repository contains the Instruct version of the 70B parameters model.** **Input** Models input text only. **Output** Models generate text only. **Model Architecture** Code Llama is an auto-regressive language model that uses an optimized transformer architecture. It was fine-tuned with up to 16k tokens. This variant **does not** support long context of up to 100k tokens. **Model Dates** Code Llama and its variants have been trained between January 2023 and January 2024. **Status** This is a static model trained on an offline dataset. Future versions of Code Llama - Instruct will be released as we improve model safety with community feedback. **License** A custom commercial license is available at: [https://ai.meta.com/resources/models-and-libraries/llama-downloads/](https://ai.meta.com/resources/models-and-libraries/llama-downloads/) **Research Paper** More information can be found in the paper "[Code Llama: Open Foundation Models for Code](https://ai.meta.com/research/publications/code-llama-open-foundation-models-for-code/)" or its [arXiv page](https://arxiv.org/abs/2308.12950). ## Intended Use **Intended Use Cases** Code Llama and its variants are intended for commercial and research use in English and relevant programming languages. The base model Code Llama can be adapted for a variety of code synthesis and understanding tasks, Code Llama - Python is designed specifically to handle the Python programming language, and Code Llama - Instruct is intended to be safer to use for code assistant and generation applications. **Out-of-Scope Uses** Use in any manner that violates applicable laws or regulations (including trade compliance laws). Use in languages other than English. Use in any other way that is prohibited by the Acceptable Use Policy and Licensing Agreement for Code Llama and its variants. ## Hardware and Software **Training Factors** We used custom training libraries. The training and fine-tuning of the released models have been performed Meta’s Research Super Cluster.\\**Carbon Footprint** In aggregate, training all 12 Code Llama models required 1400K GPU hours of computation on hardware of type A100-80GB (TDP of 350-400W). Estimated total emissions were 228.55 tCO2eq, 100% of which were offset by Meta’s sustainability program. ## Evaluation Results See evaluations for the main models and detailed ablations in Section 3 and safety evaluations in Section 4 of the research paper. ## Ethical Considerations and Limitations Code Llama and its variants are a new technology that carries risks with use. Testing conducted to date has been in English, and has not covered, nor could it cover all scenarios. For these reasons, as with all LLMs, Code Llama’s potential outputs cannot be predicted in advance, and the model may in some instances produce inaccurate or objectionable responses to user prompts. Therefore, before deploying any applications of Code Llama, developers should perform safety testing and tuning tailored to their specific applications of the model. Please see the Responsible Use Guide available available at [https://ai.meta.com/llama/responsible-use-guide](https://ai.meta.com/llama/responsible-use-guide).
HHazard/llama2-13b-qodly
HHazard
2024-01-30T13:00:59Z
6
0
transformers
[ "transformers", "tensorboard", "safetensors", "llama", "text-generation", "autotrain", "conversational", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-01-26T15:54:00Z
--- tags: - autotrain - text-generation widget: - text: "I love AutoTrain because " license: other --- # Model Trained Using AutoTrain This model was trained using AutoTrain. For more information, please visit [AutoTrain](https://hf.co/docs/autotrain). # Usage ```python from transformers import AutoModelForCausalLM, AutoTokenizer model_path = "PATH_TO_THIS_REPO" tokenizer = AutoTokenizer.from_pretrained(model_path) model = AutoModelForCausalLM.from_pretrained( model_path, device_map="auto", torch_dtype='auto' ).eval() # Prompt content: "hi" messages = [ {"role": "user", "content": "hi"} ] input_ids = tokenizer.apply_chat_template(conversation=messages, tokenize=True, add_generation_prompt=True, return_tensors='pt') output_ids = model.generate(input_ids.to('cuda')) response = tokenizer.decode(output_ids[0][input_ids.shape[1]:], skip_special_tokens=True) # Model response: "Hello! How can I assist you today?" print(response) ```
Harshit4199/ppo-LunarLander-v2
Harshit4199
2024-01-30T12:54:58Z
0
0
stable-baselines3
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2024-01-30T12:51:23Z
--- library_name: stable-baselines3 tags: - LunarLander-v2 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 metrics: - type: mean_reward value: 252.23 +/- 21.51 name: mean_reward verified: false --- # **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
birgermoell/Rapid-Cycling
birgermoell
2024-01-30T12:54:34Z
46
2
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "merge", "mergekit", "lazymergekit", "timpal0l/Mistral-7B-v0.1-flashback-v2", "RJuro/munin-neuralbeagle-7b", "base_model:RJuro/munin-neuralbeagle-7b", "base_model:merge:RJuro/munin-neuralbeagle-7b", "base_model:timpal0l/Mistral-7B-v0.1-flashback-v2", "base_model:merge:timpal0l/Mistral-7B-v0.1-flashback-v2", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-01-30T12:44:32Z
--- tags: - merge - mergekit - lazymergekit - timpal0l/Mistral-7B-v0.1-flashback-v2 - RJuro/munin-neuralbeagle-7b base_model: - timpal0l/Mistral-7B-v0.1-flashback-v2 - RJuro/munin-neuralbeagle-7b --- # Rapid-Cycling ![](https://huggingface.co/birgermoell/Rapid-Cycling/resolve/main/rapid_cycling.png?download=true) Rapid-Cycling is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing): * [timpal0l/Mistral-7B-v0.1-flashback-v2](https://huggingface.co/timpal0l/Mistral-7B-v0.1-flashback-v2) * [RJuro/munin-neuralbeagle-7b](https://huggingface.co/RJuro/munin-neuralbeagle-7b) ## 🧩 Configuration ```yaml slices: - sources: - model: timpal0l/Mistral-7B-v0.1-flashback-v2 layer_range: [0, 32] - model: RJuro/munin-neuralbeagle-7b layer_range: [0, 32] merge_method: slerp base_model: timpal0l/Mistral-7B-v0.1-flashback-v2 parameters: t: - filter: self_attn value: [0, 0.5, 0.3, 0.7, 1] - filter: mlp value: [1, 0.5, 0.7, 0.3, 0] - value: 0.5 dtype: bfloat16 ``` ## 💻 Usage ```python !pip install -qU transformers accelerate from transformers import AutoTokenizer import transformers import torch model = "birgermoell/Rapid-Cycling" messages = [{"role": "user", "content": "What is a large language model?"}] tokenizer = AutoTokenizer.from_pretrained(model) prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) pipeline = transformers.pipeline( "text-generation", model=model, torch_dtype=torch.float16, device_map="auto", ) outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95) print(outputs[0]["generated_text"]) ```
Trubnik1967/distilbert-base-uncased-finetuned-ner
Trubnik1967
2024-01-30T12:53:49Z
92
0
transformers
[ "transformers", "tensorboard", "safetensors", "distilbert", "token-classification", "generated_from_trainer", "base_model:distilbert/distilbert-base-uncased", "base_model:finetune:distilbert/distilbert-base-uncased", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2024-01-29T08:13:47Z
--- license: apache-2.0 base_model: distilbert-base-uncased tags: - generated_from_trainer metrics: - precision - recall - f1 - accuracy model-index: - name: distilbert-base-uncased-finetuned-ner results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-ner This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.0664 - Precision: 0.9323 - Recall: 0.9432 - F1: 0.9377 - Accuracy: 0.9849 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | 0.2412 | 1.0 | 878 | 0.0726 | 0.9078 | 0.9237 | 0.9157 | 0.9800 | | 0.0497 | 2.0 | 1756 | 0.0594 | 0.9266 | 0.9353 | 0.9310 | 0.9835 | | 0.0286 | 3.0 | 2634 | 0.0609 | 0.9252 | 0.9408 | 0.9329 | 0.9842 | | 0.0172 | 4.0 | 3512 | 0.0646 | 0.9298 | 0.9397 | 0.9347 | 0.9846 | | 0.0126 | 5.0 | 4390 | 0.0664 | 0.9323 | 0.9432 | 0.9377 | 0.9849 | ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.16.1 - Tokenizers 0.15.1
yotasr/Smart_Tour_GizaVersion1.01
yotasr
2024-01-30T12:44:09Z
178
0
transformers
[ "transformers", "safetensors", "vit", "image-classification", "generated_from_trainer", "dataset:imagefolder", "base_model:yotasr/Smart_TourGuide", "base_model:finetune:yotasr/Smart_TourGuide", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
image-classification
2024-01-30T11:58:25Z
--- base_model: yotasr/Smart_TourGuide tags: - generated_from_trainer datasets: - imagefolder metrics: - accuracy model-index: - name: Smart_Tour_GizaVersion1.01 results: - task: name: Image Classification type: image-classification dataset: name: imagefolder type: imagefolder config: default split: train args: default metrics: - name: Accuracy type: accuracy value: 1.0 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # Smart_Tour_GizaVersion1.01 This model is a fine-tuned version of [yotasr/Smart_TourGuide](https://huggingface.co/yotasr/Smart_TourGuide) on the imagefolder dataset. It achieves the following results on the evaluation set: - Loss: 0.0491 - Accuracy: 1.0 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 128 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 4 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | No log | 1.0 | 1 | 4.4777 | 0.0 | | No log | 2.0 | 2 | 0.0777 | 1.0 | | No log | 3.0 | 3 | 0.0510 | 1.0 | | No log | 4.0 | 4 | 0.0491 | 1.0 | ### Framework versions - Transformers 4.36.2 - Pytorch 2.1.2+cu121 - Datasets 2.16.1 - Tokenizers 0.15.0
Lumiere123/indian-fashion
Lumiere123
2024-01-30T12:42:07Z
2
0
diffusers
[ "diffusers", "safetensors", "text-to-image", "stable-diffusion", "license:creativeml-openrail-m", "autotrain_compatible", "endpoints_compatible", "diffusers:StableDiffusionPipeline", "region:us" ]
text-to-image
2024-01-30T12:35:57Z
--- license: creativeml-openrail-m tags: - text-to-image - stable-diffusion --- ### Indian-Fashion Dreambooth model trained by Lumiere123 with [TheLastBen's fast-DreamBooth](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast-DreamBooth.ipynb) notebook Test the concept via A1111 Colab [fast-Colab-A1111](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast_stable_diffusion_AUTOMATIC1111.ipynb) Sample pictures of this concept:
tanatapanun/fine-tuned-BioBARTv2-20-epochs-1024-input-320-output
tanatapanun
2024-01-30T12:39:12Z
90
0
transformers
[ "transformers", "tensorboard", "safetensors", "bart", "text2text-generation", "generated_from_trainer", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2024-01-30T11:56:54Z
--- base_model: checkpoint_global_step_200000 tags: - generated_from_trainer metrics: - rouge model-index: - name: fine-tuned-BioBARTv2-20-epochs-1024-input-320-output results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # fine-tuned-BioBARTv2-20-epochs-1024-input-320-output This model is a fine-tuned version of [checkpoint_global_step_200000](https://huggingface.co/checkpoint_global_step_200000) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.7867 - Rouge1: 0.1624 - Rouge2: 0.0352 - Rougel: 0.1185 - Rougelsum: 0.1188 - Gen Len: 37.64 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 20 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:| | No log | 1.0 | 151 | 6.3285 | 0.0632 | 0.0007 | 0.0552 | 0.0555 | 37.84 | | No log | 2.0 | 302 | 0.9332 | 0.1075 | 0.0282 | 0.0825 | 0.0825 | 62.51 | | No log | 3.0 | 453 | 0.8092 | 0.0826 | 0.0196 | 0.0629 | 0.0622 | 28.26 | | 4.0641 | 4.0 | 604 | 0.7617 | 0.1106 | 0.0346 | 0.0814 | 0.0814 | 32.19 | | 4.0641 | 5.0 | 755 | 0.7385 | 0.1359 | 0.0266 | 0.1043 | 0.1048 | 35.85 | | 4.0641 | 6.0 | 906 | 0.7296 | 0.1507 | 0.0296 | 0.1099 | 0.1112 | 45.66 | | 0.6482 | 7.0 | 1057 | 0.7225 | 0.1315 | 0.026 | 0.0978 | 0.0992 | 36.35 | | 0.6482 | 8.0 | 1208 | 0.7165 | 0.1573 | 0.0302 | 0.1218 | 0.1222 | 42.68 | | 0.6482 | 9.0 | 1359 | 0.7191 | 0.1445 | 0.0307 | 0.1155 | 0.1156 | 30.12 | | 0.4567 | 10.0 | 1510 | 0.7281 | 0.1827 | 0.0423 | 0.1403 | 0.1408 | 47.87 | | 0.4567 | 11.0 | 1661 | 0.7320 | 0.1603 | 0.0311 | 0.1193 | 0.1193 | 33.69 | | 0.4567 | 12.0 | 1812 | 0.7395 | 0.1697 | 0.0357 | 0.1267 | 0.1267 | 46.9 | | 0.4567 | 13.0 | 1963 | 0.7515 | 0.1442 | 0.0297 | 0.1064 | 0.1065 | 31.28 | | 0.3275 | 14.0 | 2114 | 0.7549 | 0.1767 | 0.0306 | 0.1255 | 0.1259 | 49.23 | | 0.3275 | 15.0 | 2265 | 0.7680 | 0.1475 | 0.0327 | 0.1054 | 0.1057 | 37.28 | | 0.3275 | 16.0 | 2416 | 0.7760 | 0.1525 | 0.0337 | 0.1065 | 0.1072 | 38.87 | | 0.2407 | 17.0 | 2567 | 0.7797 | 0.1543 | 0.039 | 0.1163 | 0.1168 | 38.67 | | 0.2407 | 18.0 | 2718 | 0.7845 | 0.1794 | 0.0382 | 0.13 | 0.1305 | 38.47 | | 0.2407 | 19.0 | 2869 | 0.7860 | 0.1645 | 0.0372 | 0.1218 | 0.1219 | 36.26 | | 0.1982 | 20.0 | 3020 | 0.7867 | 0.1624 | 0.0352 | 0.1185 | 0.1188 | 37.64 | ### Framework versions - Transformers 4.36.2 - Pytorch 1.12.1+cu113 - Datasets 2.16.1 - Tokenizers 0.15.1
erens/GuiltyGearStriveStyleV2
erens
2024-01-30T12:33:49Z
1
1
diffusers
[ "diffusers", "safetensors", "text-to-image", "stable-diffusion", "license:creativeml-openrail-m", "autotrain_compatible", "endpoints_compatible", "diffusers:StableDiffusionPipeline", "region:us" ]
text-to-image
2024-01-29T22:30:37Z
--- license: creativeml-openrail-m tags: - text-to-image - stable-diffusion --- ### guiltygear Dreambooth model trained by erens with [TheLastBen's fast-DreamBooth](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast-DreamBooth.ipynb) notebook Test the concept via A1111 Colab [fast-Colab-A1111](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast_stable_diffusion_AUTOMATIC1111.ipynb) Sample pictures of this concept: ![image/png](https://cdn-uploads.huggingface.co/production/uploads/635cf1664fabde0df7409c28/kPh1_fwpunfo8y8hv7Fp2.png)
CesarLeblanc/geoplantbert_text_classification_model
CesarLeblanc
2024-01-30T12:22:16Z
95
0
transformers
[ "transformers", "safetensors", "bert", "text-classification", "generated_from_trainer", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2023-12-03T14:59:27Z
--- tags: - generated_from_trainer metrics: - accuracy model-index: - name: geoplantbert_text_classification_model results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # geoplantbert_text_classification_model This model was trained from scratch on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 4.2249 - Accuracy: 0.0948 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 123 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:------:|:---------------:|:--------:| | 4.1776 | 1.0 | 68400 | 4.2204 | 0.0948 | | 4.2058 | 2.0 | 136800 | 4.2118 | 0.0948 | | 4.1949 | 3.0 | 205200 | 4.2219 | 0.0948 | | 4.1297 | 4.0 | 273600 | 4.2298 | 0.0948 | | 4.2056 | 5.0 | 342000 | 4.2249 | 0.0948 | ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.1+cu121 - Datasets 2.15.0 - Tokenizers 0.15.0
ryusangwon/kobart-base-v2-159-korean
ryusangwon
2024-01-30T12:21:45Z
91
0
transformers
[ "transformers", "safetensors", "bart", "text2text-generation", "generated_from_trainer", "base_model:gogamza/kobart-base-v2", "base_model:finetune:gogamza/kobart-base-v2", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2024-01-30T06:50:02Z
--- license: mit base_model: gogamza/kobart-base-v2 tags: - generated_from_trainer metrics: - rouge model-index: - name: kobart-base-v2-159-korean results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # kobart-base-v2-159-korean This model is a fine-tuned version of [gogamza/kobart-base-v2](https://huggingface.co/gogamza/kobart-base-v2) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.2628 - Rouge1: 0.3634 - Rouge2: 0.1451 - Rougel: 0.3566 - Rougelsum: 0.3563 - Gen Len: 20.0 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - gradient_accumulation_steps: 16 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - num_epochs: 10 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:| | 0.2804 | 1.44 | 500 | 0.2805 | 0.3321 | 0.1273 | 0.3272 | 0.3276 | 19.9992 | | 0.2141 | 2.88 | 1000 | 0.2472 | 0.3577 | 0.1381 | 0.3526 | 0.3525 | 20.0 | | 0.1407 | 4.33 | 1500 | 0.2495 | 0.3615 | 0.1457 | 0.3543 | 0.3543 | 20.0 | | 0.1206 | 5.77 | 2000 | 0.2508 | 0.3592 | 0.1448 | 0.3533 | 0.3532 | 20.0 | | 0.0853 | 7.21 | 2500 | 0.2603 | 0.3623 | 0.147 | 0.3561 | 0.3562 | 20.0 | | 0.0777 | 8.65 | 3000 | 0.2628 | 0.3634 | 0.1451 | 0.3566 | 0.3563 | 20.0 | ### Framework versions - Transformers 4.36.2 - Pytorch 2.1.2+cu121 - Datasets 2.15.0 - Tokenizers 0.15.0
Mukalingam0813/multilingual-Distilbert-intent-classification
Mukalingam0813
2024-01-30T12:21:17Z
92
0
transformers
[ "transformers", "safetensors", "distilbert", "text-classification", "generated_from_trainer", "base_model:distilbert/distilbert-base-multilingual-cased", "base_model:finetune:distilbert/distilbert-base-multilingual-cased", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-01-30T09:20:17Z
--- license: apache-2.0 base_model: distilbert-base-multilingual-cased tags: - generated_from_trainer metrics: - accuracy model-index: - name: multilingual-Distilbert-intent-classification results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # multilingual-Distilbert-intent-classification This model is a fine-tuned version of [distilbert-base-multilingual-cased](https://huggingface.co/distilbert-base-multilingual-cased) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.1198 - Accuracy: 0.9797 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:------:|:---------------:|:--------:| | 0.1211 | 1.0 | 77839 | 0.1477 | 0.9701 | | 0.0732 | 2.0 | 155678 | 0.1170 | 0.9776 | | 0.0354 | 3.0 | 233517 | 0.1198 | 0.9797 | ### Framework versions - Transformers 4.36.2 - Pytorch 2.1.1+cu118 - Datasets 2.16.1 - Tokenizers 0.15.0
Lalith16/Zephyr7B-10epoch-CC_dataset
Lalith16
2024-01-30T12:14:07Z
0
0
null
[ "tensorboard", "safetensors", "trl", "sft", "generated_from_trainer", "base_model:HuggingFaceH4/zephyr-7b-beta", "base_model:finetune:HuggingFaceH4/zephyr-7b-beta", "license:mit", "region:us" ]
null
2024-01-30T12:13:43Z
--- license: mit base_model: HuggingFaceH4/zephyr-7b-beta tags: - trl - sft - generated_from_trainer model-index: - name: results results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # results This model is a fine-tuned version of [HuggingFaceH4/zephyr-7b-beta](https://huggingface.co/HuggingFaceH4/zephyr-7b-beta) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.7065 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.00025 - train_batch_size: 4 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: constant - lr_scheduler_warmup_ratio: 0.03 - num_epochs: 10 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 0.7779 | 0.35 | 100 | 1.0474 | | 0.6553 | 0.69 | 200 | 0.8922 | | 0.6339 | 1.04 | 300 | 0.7221 | | 0.6018 | 1.39 | 400 | 0.7020 | | 0.5853 | 1.74 | 500 | 0.6908 | | 0.445 | 2.08 | 600 | 0.6887 | | 0.4875 | 2.43 | 700 | 0.6783 | | 0.4938 | 2.78 | 800 | 0.6883 | | 0.3598 | 3.12 | 900 | 0.6893 | | 0.3549 | 3.47 | 1000 | 0.6763 | | 0.3624 | 3.82 | 1100 | 0.6971 | | 0.344 | 4.17 | 1200 | 0.7348 | | 0.3393 | 4.51 | 1300 | 0.7502 | | 0.3823 | 4.86 | 1400 | 0.6861 | | 0.3534 | 5.21 | 1500 | 0.6849 | | 0.3632 | 5.56 | 1600 | 0.6585 | | 0.3634 | 5.9 | 1700 | 0.6414 | | 0.3002 | 6.25 | 1800 | 0.6662 | | 0.3126 | 6.6 | 1900 | 0.6864 | | 0.3129 | 6.94 | 2000 | 0.6638 | | 0.259 | 7.29 | 2100 | 0.6967 | | 0.27 | 7.64 | 2200 | 0.7059 | | 0.3063 | 7.99 | 2300 | 0.6582 | | 0.2814 | 8.33 | 2400 | 0.7205 | | 0.3005 | 8.68 | 2500 | 0.7334 | | 0.2862 | 9.03 | 2600 | 0.6839 | | 0.3092 | 9.38 | 2700 | 0.6929 | | 0.3078 | 9.72 | 2800 | 0.7065 | ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.16.1 - Tokenizers 0.15.1
TvERn/my-pet
TvERn
2024-01-30T12:13:35Z
2
0
diffusers
[ "diffusers", "safetensors", "NxtWave-GenAI-Webinar", "text-to-image", "stable-diffusion", "license:creativeml-openrail-m", "autotrain_compatible", "endpoints_compatible", "diffusers:StableDiffusionPipeline", "region:us" ]
text-to-image
2024-01-30T12:06:32Z
--- license: creativeml-openrail-m tags: - NxtWave-GenAI-Webinar - text-to-image - stable-diffusion --- ### My-Pet- Dreambooth model trained by TvERn following the "Build your own Gen AI model" session by NxtWave. Project Submission Code: A16 Sample pictures of this concept:
FelixChao/Patronum-7B
FelixChao
2024-01-30T12:10:58Z
81
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "arxiv:1910.09700", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-01-30T12:05:06Z
--- license: apache-2.0 --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> This modelcard aims to be a base template for new models. It has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/modelcard_template.md?plain=1). ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
hiraltalsaniya/llama2-7b-fine-tune-company-task-classification
hiraltalsaniya
2024-01-30T12:03:14Z
23
0
transformers
[ "transformers", "pytorch", "llama", "text-generation", "facebook", "meta", "llama-2", "text-classification", "en", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-classification
2024-01-23T05:51:25Z
--- language: - en pipeline_tag: text-classification tags: - facebook - meta - pytorch - llama - llama-2 license: mit ---
golesheed/whisper-non-native-children-7-dutch
golesheed
2024-01-30T12:02:18Z
56
0
transformers
[ "transformers", "tensorboard", "safetensors", "whisper", "automatic-speech-recognition", "generated_from_trainer", "nl", "base_model:openai/whisper-large-v2", "base_model:finetune:openai/whisper-large-v2", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2024-01-30T09:52:07Z
--- language: - nl license: apache-2.0 base_model: openai/whisper-large-v2 tags: - generated_from_trainer metrics: - wer model-index: - name: Whisper Large V2 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # Whisper Large V2 This model is a fine-tuned version of [openai/whisper-large-v2](https://huggingface.co/openai/whisper-large-v2) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.4038 - Wer: 14.0551 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 3e-05 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 20 - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:-------:| | 0.672 | 0.71 | 30 | 0.3839 | 16.2013 | | 0.2682 | 1.43 | 60 | 0.3620 | 13.6562 | | 0.1681 | 2.14 | 90 | 0.3700 | 14.9478 | | 0.0726 | 2.86 | 120 | 0.3728 | 13.3713 | | 0.0429 | 3.57 | 150 | 0.3946 | 14.5109 | | 0.0223 | 4.29 | 180 | 0.3921 | 14.2640 | | 0.0114 | 5.0 | 210 | 0.4038 | 14.0551 | ### Framework versions - Transformers 4.38.0.dev0 - Pytorch 2.1.0+cu121 - Datasets 2.14.6 - Tokenizers 0.15.0
balaramas/bart-large-summariser-finetune
balaramas
2024-01-30T11:55:56Z
155
0
transformers
[ "transformers", "pytorch", "tf", "jax", "rust", "safetensors", "bart", "text2text-generation", "summarization", "en", "dataset:cnn_dailymail", "arxiv:1910.13461", "license:mit", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
summarization
2024-01-30T11:45:36Z
--- language: - en tags: - summarization license: mit thumbnail: https://huggingface.co/front/thumbnails/facebook.png datasets: - cnn_dailymail model-index: - name: facebook/bart-large-cnn results: - task: type: summarization name: Summarization dataset: name: cnn_dailymail type: cnn_dailymail config: 3.0.0 split: train metrics: - name: ROUGE-1 type: rouge value: 42.9486 verified: true - name: ROUGE-2 type: rouge value: 20.8149 verified: true - name: ROUGE-L type: rouge value: 30.6186 verified: true - name: ROUGE-LSUM type: rouge value: 40.0376 verified: true - name: loss type: loss value: 2.529000997543335 verified: true - name: gen_len type: gen_len value: 78.5866 verified: true --- # BART (large-sized model), fine-tuned on CNN Daily Mail BART model pre-trained on English language, and fine-tuned on [CNN Daily Mail](https://huggingface.co/datasets/cnn_dailymail). It was introduced in the paper [BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension](https://arxiv.org/abs/1910.13461) by Lewis et al. and first released in [this repository (https://github.com/pytorch/fairseq/tree/master/examples/bart). Disclaimer: The team releasing BART did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description BART is a transformer encoder-encoder (seq2seq) model with a bidirectional (BERT-like) encoder and an autoregressive (GPT-like) decoder. BART is pre-trained by (1) corrupting text with an arbitrary noising function, and (2) learning a model to reconstruct the original text. BART is particularly effective when fine-tuned for text generation (e.g. summarization, translation) but also works well for comprehension tasks (e.g. text classification, question answering). This particular checkpoint has been fine-tuned on CNN Daily Mail, a large collection of text-summary pairs. ## Intended uses & limitations You can use this model for text summarization. ### How to use Here is how to use this model with the [pipeline API](https://huggingface.co/transformers/main_classes/pipelines.html): ```python from transformers import pipeline summarizer = pipeline("summarization", model="facebook/bart-large-cnn") ARTICLE = """ New York (CNN)When Liana Barrientos was 23 years old, she got married in Westchester County, New York. A year later, she got married again in Westchester County, but to a different man and without divorcing her first husband. Only 18 days after that marriage, she got hitched yet again. Then, Barrientos declared "I do" five more times, sometimes only within two weeks of each other. In 2010, she married once more, this time in the Bronx. In an application for a marriage license, she stated it was her "first and only" marriage. Barrientos, now 39, is facing two criminal counts of "offering a false instrument for filing in the first degree," referring to her false statements on the 2010 marriage license application, according to court documents. Prosecutors said the marriages were part of an immigration scam. On Friday, she pleaded not guilty at State Supreme Court in the Bronx, according to her attorney, Christopher Wright, who declined to comment further. After leaving court, Barrientos was arrested and charged with theft of service and criminal trespass for allegedly sneaking into the New York subway through an emergency exit, said Detective Annette Markowski, a police spokeswoman. In total, Barrientos has been married 10 times, with nine of her marriages occurring between 1999 and 2002. All occurred either in Westchester County, Long Island, New Jersey or the Bronx. She is believed to still be married to four men, and at one time, she was married to eight men at once, prosecutors say. Prosecutors said the immigration scam involved some of her husbands, who filed for permanent residence status shortly after the marriages. Any divorces happened only after such filings were approved. It was unclear whether any of the men will be prosecuted. The case was referred to the Bronx District Attorney\'s Office by Immigration and Customs Enforcement and the Department of Homeland Security\'s Investigation Division. Seven of the men are from so-called "red-flagged" countries, including Egypt, Turkey, Georgia, Pakistan and Mali. Her eighth husband, Rashid Rajput, was deported in 2006 to his native Pakistan after an investigation by the Joint Terrorism Task Force. If convicted, Barrientos faces up to four years in prison. Her next court appearance is scheduled for May 18. """ print(summarizer(ARTICLE, max_length=130, min_length=30, do_sample=False)) >>> [{'summary_text': 'Liana Barrientos, 39, is charged with two counts of "offering a false instrument for filing in the first degree" In total, she has been married 10 times, with nine of her marriages occurring between 1999 and 2002. She is believed to still be married to four men.'}] ``` ### BibTeX entry and citation info ```bibtex @article{DBLP:journals/corr/abs-1910-13461, author = {Mike Lewis and Yinhan Liu and Naman Goyal and Marjan Ghazvininejad and Abdelrahman Mohamed and Omer Levy and Veselin Stoyanov and Luke Zettlemoyer}, title = {{BART:} Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension}, journal = {CoRR}, volume = {abs/1910.13461}, year = {2019}, url = {http://arxiv.org/abs/1910.13461}, eprinttype = {arXiv}, eprint = {1910.13461}, timestamp = {Thu, 31 Oct 2019 14:02:26 +0100}, biburl = {https://dblp.org/rec/journals/corr/abs-1910-13461.bib}, bibsource = {dblp computer science bibliography, https://dblp.org} } ```
mertbozkurt/Llama-2-7b-TR-recipe
mertbozkurt
2024-01-30T11:47:34Z
6
1
transformers
[ "transformers", "pytorch", "llama", "text-generation", "dataset:mertbozkurt/turkish-recipe", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2023-12-24T16:24:27Z
--- license: mit datasets: - mertbozkurt/turkish-recipe --- # This model is a fine-tuned version of [NousResearch/Llama-2-7b-chat-hf](https://huggingface.co/NousResearch/Llama-2-7b-chat-hf) on [mertbozkurt/turkish-recipe](https://huggingface.co/datasets/mertbozkurt/llama2-TR-recipe) dataset. Test example [Notebook](https://github.com/mertorelio/llm-apps/blob/main/Fine_tune_Llama_2.ipynb)
Kittuu/peacock
Kittuu
2024-01-30T11:46:27Z
0
0
diffusers
[ "diffusers", "safetensors", "NxtWave-GenAI-Webinar", "text-to-image", "stable-diffusion", "license:creativeml-openrail-m", "autotrain_compatible", "endpoints_compatible", "diffusers:StableDiffusionPipeline", "region:us" ]
text-to-image
2024-01-30T11:42:19Z
--- license: creativeml-openrail-m tags: - NxtWave-GenAI-Webinar - text-to-image - stable-diffusion --- ### Peacock Dreambooth model trained by Kittuu following the "Build your own Gen AI model" session by NxtWave. Project Submission Code: 4MW21CS038 Sample pictures of this concept:
ZiHDeng/peft-lora-starcoder1B-Instruction-ny8-MIX-2000
ZiHDeng
2024-01-30T11:45:46Z
4
0
peft
[ "peft", "tensorboard", "safetensors", "generated_from_trainer", "base_model:bigcode/starcoderbase-1b", "base_model:adapter:bigcode/starcoderbase-1b", "license:bigcode-openrail-m", "region:us" ]
null
2024-01-30T09:33:10Z
--- license: bigcode-openrail-m library_name: peft tags: - generated_from_trainer base_model: bigcode/starcoderbase-1b model-index: - name: peft-lora-starcoder1B-Instruction-ny8-MIX-2000 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # peft-lora-starcoder1B-Instruction-ny8-MIX-2000 This model is a fine-tuned version of [bigcode/starcoderbase-1b](https://huggingface.co/bigcode/starcoderbase-1b) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.1039 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0005 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 16 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 30 - training_steps: 2000 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 0.1705 | 0.05 | 100 | 0.1413 | | 0.1338 | 0.1 | 200 | 0.1216 | | 0.1191 | 0.15 | 300 | 0.1143 | | 0.1147 | 0.2 | 400 | 0.1084 | | 0.1109 | 0.25 | 500 | 0.1063 | | 0.1061 | 0.3 | 600 | 0.1041 | | 0.1059 | 0.35 | 700 | 0.1037 | | 0.1018 | 0.4 | 800 | 0.1026 | | 0.1011 | 0.45 | 900 | 0.1021 | | 0.0996 | 0.5 | 1000 | 0.1017 | | 0.0981 | 0.55 | 1100 | 0.1010 | | 0.0961 | 0.6 | 1200 | 0.1012 | | 0.0941 | 0.65 | 1300 | 0.1020 | | 0.0926 | 0.7 | 1400 | 0.1018 | | 0.0907 | 0.75 | 1500 | 0.1026 | | 0.0901 | 0.8 | 1600 | 0.1028 | | 0.0872 | 0.85 | 1700 | 0.1034 | | 0.0889 | 0.9 | 1800 | 0.1037 | | 0.087 | 0.95 | 1900 | 0.1038 | | 0.0876 | 1.0 | 2000 | 0.1039 | ### Framework versions - PEFT 0.7.1 - Transformers 4.36.2 - Pytorch 2.1.2+cu121 - Datasets 2.16.1 - Tokenizers 0.15.0
Wajid333/Reinforce
Wajid333
2024-01-30T11:41:52Z
0
0
null
[ "CartPole-v1", "reinforce", "reinforcement-learning", "custom-implementation", "deep-rl-class", "model-index", "region:us" ]
reinforcement-learning
2024-01-30T11:41:43Z
--- tags: - CartPole-v1 - reinforce - reinforcement-learning - custom-implementation - deep-rl-class model-index: - name: Reinforce results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: CartPole-v1 type: CartPole-v1 metrics: - type: mean_reward value: 500.00 +/- 0.00 name: mean_reward verified: false --- # **Reinforce** Agent playing **CartPole-v1** This is a trained model of a **Reinforce** agent playing **CartPole-v1** . To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: https://huggingface.co/deep-rl-course/unit4/introduction
sadhaklal/mlp-cifar2
sadhaklal
2024-01-30T11:37:54Z
0
0
pytorch
[ "pytorch", "image-classification", "dataset:cifar10", "region:us" ]
image-classification
2024-01-29T17:03:43Z
--- datasets: - cifar10 metrics: - accuracy pipeline_tag: image-classification library_name: pytorch --- # mlp-cifar2 Multi-layer perceptron (MLP) trained on CIFAR-2 (a subset of CIFAR-10 for classifying 'airplane' vs. 'bird'). `RandomCrop(24)` was applied on the training set images, and `Resize(24)` was applied on the validation set images. This model pertains to Exercise 1 of Chapter 7 of the book "Deep Learning with PyTorch" by Eli Stevens, Luca Antiga, and Thomas Viehmann. Code: https://github.com/sambitmukherjee/dlwpt-exercises/blob/main/chapter_7/exercise_1.ipynb Experiment tracking: https://wandb.ai/sadhaklal/mlp-cifar2 ## Usage ``` !pip install -q datasets from datasets import load_dataset cifar10 = load_dataset("cifar10") label_map = {0: 0, 2: 1} class_names = ['airplane', 'bird'] cifar2_train = [(example['img'], label_map[example['label']]) for example in cifar10['train'] if example['label'] in [0, 2]] cifar2_val = [(example['img'], label_map[example['label']]) for example in cifar10['test'] if example['label'] in [0, 2]] example = cifar2_val[0] img, label = example import torch from torchvision.transforms import v2 val_tfms = v2.Compose([ v2.Resize(24), v2.ToImage(), v2.ToDtype(torch.float32, scale=True), v2.Normalize(mean=[0.4915, 0.4823, 0.4468], std=[0.2470, 0.2435, 0.2616]) ]) img = val_tfms(img) batch = img.reshape(-1).unsqueeze(0) # Flatten. import torch.nn as nn from huggingface_hub import PyTorchModelHubMixin class MLPForCIFAR2(nn.Module, PyTorchModelHubMixin): """Multi-layer perceptron (MLP) for classifying 'airplane' vs. 'bird' in the CIFAR-2 dataset (a subset of CIFAR-10).""" def __init__(self): super().__init__() self.mlp = nn.Sequential( nn.Linear(1728, 1024), # Hidden layer. nn.Tanh(), nn.Linear(1024, 512), # Hidden layer. nn.Tanh(), nn.Linear(512, 128), # Hidden layer. nn.Tanh(), nn.Linear(128, 2) # Output layer. ) def forward(self, x): return self.mlp(x) model = MLPForCIFAR2.from_pretrained("sadhaklal/mlp-cifar2") model.eval() import torch.nn.functional as F with torch.no_grad(): logits = model(batch) pred = logits[0].argmax().item() proba = F.softmax(logits, dim=1) print(f"Predicted class: {class_names[pred]}") print(f"Predicted class probabilities ('airplane' vs. 'bird'): {proba[0].tolist()}") ``` ## Metric Accuracy on `cifar2_val`: 0.8090
vilm/Mixsmol-4x400M-v0.1-epoch2
vilm
2024-01-30T11:33:30Z
174
5
transformers
[ "transformers", "safetensors", "mixtral", "text-generation", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-01-28T03:04:11Z
--- license: apache-2.0 widget: - text: My name is El Microondas the Wise, and example_title: El Microondas - text: Kennesaw State University is a public example_title: Kennesaw State University - text: Bungie Studios is an American video game developer. They are most famous for developing the award winning Halo series of video games. They also made Destiny. The studio was founded example_title: Bungie - text: The Mona Lisa is a world-renowned painting created by example_title: Mona Lisa - text: The Harry Potter series, written by J.K. Rowling, begins with the book titled example_title: Harry Potter Series - text: 'Question: I have cities, but no houses. I have mountains, but no trees. I have water, but no fish. What am I? Answer:' example_title: Riddle - text: The process of photosynthesis involves the conversion of example_title: Photosynthesis - text: Jane went to the store to buy some groceries. She picked up apples, oranges, and a loaf of bread. When she got home, she realized she forgot example_title: Story Continuation - text: 'Problem 2: If a train leaves Station A at 9:00 AM and travels at 60 mph, and another train leaves Station B at 10:00 AM and travels at 80 mph, when will they meet if the distance between the stations is 300 miles? To determine' example_title: Math Problem - text: In the context of computer programming, an algorithm is example_title: Algorithm Definition --- # Mixsmol-4x400M-v0.1 by Ontocord This is the first checkpoint (Epoch 1) of Mixsmol-4x400M-v0.1 Note that this is an experimental in data mixing. Therefore, we only trained the model on 50B tokens (95% English and 5% Vietnamese) to test the following: - Reasoining capabilities through high-quality synthetic textbooks data pretraining - Crosslingual understanding through machine translation and multilingual + multiple tasks pretraining After verifying our hypothesis with this run, we will schedule a second run on bigger data and compute for it to achieve its maximum capability. ## Data - Synthetic Textbooks: 8M samples - RefinedWeb: 1M samples - RedPajama-v2: 500K samples - MathPile: Everything - ThePile: MiniPile Subset - GoodWiki - The Stack Smol XL - The Vault: train_small split - Instruction Pretraining: 250k samples
firdho26/wav2vec2-large-xlsr-53-english-finetuned-ravdess
firdho26
2024-01-30T11:33:23Z
16
0
transformers
[ "transformers", "tensorboard", "safetensors", "wav2vec2", "audio-classification", "generated_from_trainer", "dataset:narad/ravdess", "base_model:jonatasgrosman/wav2vec2-large-xlsr-53-english", "base_model:finetune:jonatasgrosman/wav2vec2-large-xlsr-53-english", "license:apache-2.0", "model-index", "endpoints_compatible", "region:us" ]
audio-classification
2024-01-30T07:47:08Z
--- license: apache-2.0 base_model: jonatasgrosman/wav2vec2-large-xlsr-53-english tags: - generated_from_trainer datasets: - narad/ravdess metrics: - accuracy - precision - recall - f1 model-index: - name: wav2vec2-large-xlsr-53-english-finetuned-ravdess results: - task: name: Audio Classification type: audio-classification dataset: name: RAVDESS type: narad/ravdess config: all split: train args: all metrics: - name: Accuracy type: accuracy value: 0.8298611111111112 - name: Precision type: precision value: 0.8453025128787324 - name: Recall type: recall value: 0.8298611111111112 - name: F1 type: f1 value: 0.8329568451751053 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # wav2vec2-large-xlsr-53-english-finetuned-ravdess This model is a fine-tuned version of [jonatasgrosman/wav2vec2-large-xlsr-53-english](https://huggingface.co/jonatasgrosman/wav2vec2-large-xlsr-53-english) on the RAVDESS dataset. It achieves the following results on the evaluation set: - Loss: 0.5624 - Accuracy: 0.8299 - Precision: 0.8453 - Recall: 0.8299 - F1: 0.8330 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 2 - eval_batch_size: 2 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 4 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 6 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 | |:-------------:|:-----:|:----:|:---------------:|:--------:|:---------:|:------:|:------:| | 1.9765 | 1.0 | 288 | 1.9102 | 0.3090 | 0.3203 | 0.3090 | 0.1941 | | 1.4803 | 2.0 | 576 | 1.4590 | 0.5660 | 0.5493 | 0.5660 | 0.4811 | | 1.1625 | 3.0 | 864 | 1.2308 | 0.6215 | 0.6299 | 0.6215 | 0.5936 | | 0.8354 | 4.0 | 1152 | 0.7821 | 0.7222 | 0.7555 | 0.7222 | 0.6869 | | 0.2066 | 5.0 | 1440 | 0.7910 | 0.7708 | 0.8373 | 0.7708 | 0.7881 | | 0.6335 | 6.0 | 1728 | 0.5624 | 0.8299 | 0.8453 | 0.8299 | 0.8330 | ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.16.1 - Tokenizers 0.15.1