modelId
stringlengths
5
139
author
stringlengths
2
42
last_modified
timestamp[us, tz=UTC]date
2020-02-15 11:33:14
2025-09-03 06:27:42
downloads
int64
0
223M
likes
int64
0
11.7k
library_name
stringclasses
535 values
tags
listlengths
1
4.05k
pipeline_tag
stringclasses
55 values
createdAt
timestamp[us, tz=UTC]date
2022-03-02 23:29:04
2025-09-03 06:27:02
card
stringlengths
11
1.01M
dragonSwing/viwav2vec2-base-3k
dragonSwing
2023-07-04T07:01:32Z
176
2
transformers
[ "transformers", "pytorch", "safetensors", "wav2vec2", "pretraining", "speech", "automatic-speech-recognition", "vi", "arxiv:2006.11477", "license:cc-by-sa-4.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-05-03T01:16:58Z
--- license: cc-by-sa-4.0 language: vi tags: - speech - automatic-speech-recognition --- # Wav2Vec2 base model trained of 3K hours of Vietnamese speech The base model is pre-trained on 16kHz sampled speech audio from Vietnamese speech corpus containing 3K hours of spontaneous, reading, and broadcasting speech. When using the model make sure that your speech input is also sampled at 16Khz. Note that this model should be fine-tuned on a downstream task, like Vietnamese Automatic Speech Recognition. **Note**: This model does not have a tokenizer as it was pretrained on audio alone. In order to use this model speech recognition, a tokenizer should be created and the model should be fine-tuned on labeled text data. Check out [this blog](https://huggingface.co/blog/fine-tune-wav2vec2-english) for more in-detail explanation of how to fine-tune the model. [Facebook's Wav2Vec2 blog](https://ai.facebook.com/blog/wav2vec-20-learning-the-structure-of-speech-from-raw-audio/) [Paper](https://arxiv.org/abs/2006.11477) # Usage See [this notebook](https://colab.research.google.com/drive/1FjTsqbYKphl9kL-eILgUc-bl4zVThL8F?usp=sharing) for more information on how to fine-tune the English pre-trained model. ```python import torch from transformers import Wav2Vec2Model model = Wav2Vec2Model.from_pretrained("dragonSwing/viwav2vec2-base-3k") # Sanity check inputs = torch.rand([1, 16000]) outputs = model(inputs) ```
megagonlabs/pilota_dialog
megagonlabs
2023-07-04T06:57:39Z
0
0
null
[ "t5", "text2text-generation", "pilota", "ja", "license:apache-2.0", "region:us" ]
text2text-generation
2023-06-27T07:33:45Z
--- language: ja tags: - t5 - text2text-generation - pilota license: apache-2.0 --- # Pilota model for dialogs A model for [Pilota](https://github.com/megagonlabs/pilota) trained with [Accommodation Search Dialog Corpus](https://github.com/megagonlabs/asdc) and other additional examples - ``scud`` - Fine tuned model of [t5-base-japanese-web (with Byte-fallback, 8K)](https://huggingface.co/megagonlabs/t5-base-japanese-web-8k) - The original model is distributed in [the Apache License 2.0](https://www.apache.org/licenses/LICENSE-2.0) - ``scorer`` - Fine tuned model of [LINE DistilBERT Japanese](https://huggingface.co/line-corporation/line-distilbert-base-japanese) - The original model is distributed in [the Apache License 2.0](https://www.apache.org/licenses/LICENSE-2.0) ## Usage 1. Install [Pilota](https://github.com/megagonlabs/pilota) 2. Prepare inputs - Command ```bash echo -e 'ใ”่ฆๆœ›ใ‚’ใŠ็Ÿฅใ‚‰ใ›ใใ ใ•ใ„\tใฏใ„ใ€‚้ƒจๅฑ‹ใ‹ใ‚‰ๅฏŒๅฃซๅฑฑใŒ่ฆ‹ใˆใฆใ€ๅคœๆ™ฏใ‚’่ฆ‹ใชใŒใ‚‰้ฃŸไบ‹ใฎใงใใ‚‹ใƒ›ใƒ†ใƒซใŒใ„ใ„ใชใ€‚\nใ“ใ‚“ใซใกใฏ\tใ“ใ‚“ใซใกใฏ' | python -m pilota.convert.plain2request | tee input.jsonl ``` - Output ```jsonl {"context": [{"name": "agent", "text": "ใ”่ฆๆœ›ใ‚’ใŠ็Ÿฅใ‚‰ใ›ใใ ใ•ใ„"}], "utterance": "ใฏใ„ใ€‚้ƒจๅฑ‹ใ‹ใ‚‰ๅฏŒๅฃซๅฑฑใŒ่ฆ‹ใˆใฆใ€ๅคœๆ™ฏใ‚’่ฆ‹ใชใŒใ‚‰้ฃŸไบ‹ใฎใงใใ‚‹ใƒ›ใƒ†ใƒซใŒใ„ใ„ใชใ€‚", "sentences": null, "meta": {}} {"context": [{"name": "agent", "text": "ใ“ใ‚“ใซใกใฏ"}], "utterance": "ใ“ใ‚“ใซใกใฏ", "sentences": null, "meta": {}} ``` 3. Feed it to Pilota - Command ```console pilota -m megagonlabs/pilota_dialog --batch_size 1 --outlen 60 --nbest 1 --beam 5 < input.jsonl ``` - Output ```jsonl [{"scuds_nbest": [[]], "original_ranks": [0], "scores": [0.9911208689212798], "scores_detail": [{"OK": 0.9704028964042664, "incorrect_none": 0.04205145686864853, "lack": 0.0007874675211496651, "limited": 0.0003119863977190107, "non_fluent": 0.0002362923405598849, "untruth": 0.0013080810895189643}], "sentence": "ใฏใ„ใ€‚"}, {"scuds_nbest": [["้ƒจๅฑ‹ใ‹ใ‚‰ๅฏŒๅฃซๅฑฑใŒ่ฆ‹ใˆใ‚‹ใƒ›ใƒ†ใƒซใŒ่‰ฏใ„ใ€‚", "ๅคœๆ™ฏใ‚’่ฆ‹ใชใŒใ‚‰้ฃŸไบ‹ใฎใงใใ‚‹ใƒ›ใƒ†ใƒซใŒ่‰ฏใ„ใ€‚"]], "original_ranks": [0], "scores": [0.9952289938926696], "scores_detail": [{"OK": 0.9840966463088989, "incorrect_none": 0.010280555114150047, "lack": 0.0032871251460164785, "limited": 0.00041511686868034303, "non_fluent": 0.0002954243100248277, "untruth": 0.003289491171017289}], "sentence": "้ƒจๅฑ‹ใ‹ใ‚‰ๅฏŒๅฃซๅฑฑใŒ่ฆ‹ใˆใฆใ€ๅคœๆ™ฏใ‚’่ฆ‹ใชใŒใ‚‰้ฃŸไบ‹ใฎใงใใ‚‹ใƒ›ใƒ†ใƒซใŒใ„ใ„ใชใ€‚"}] [{"scuds_nbest": [[]], "original_ranks": [0], "scores": [0.9831213414669036], "scores_detail": [{"OK": 0.9704028964042664, "incorrect_none": 0.04205145686864853, "lack": 0.0007874675211496651, "limited": 0.0003119863977190107, "non_fluent": 0.0002362923405598849, "untruth": 0.0013080810895189643}], "sentence": "ใ“ใ‚“ใซใกใฏ"}] ``` ## License Apache License 2.0
Arindam75/Pyramids
Arindam75
2023-07-04T06:47:35Z
1
0
ml-agents
[ "ml-agents", "tensorboard", "onnx", "Pyramids", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Pyramids", "region:us" ]
reinforcement-learning
2023-07-04T06:47:34Z
--- library_name: ml-agents tags: - Pyramids - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Pyramids --- # **ppo** Agent playing **Pyramids** This is a trained model of a **ppo** agent playing **Pyramids** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog ๐Ÿถ to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: Arindam75/Pyramids 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play ๐Ÿ‘€
NasimB/gpt2-dp-cl-rarity-8-276k
NasimB
2023-07-04T06:15:41Z
4
0
transformers
[ "transformers", "pytorch", "gpt2", "text-generation", "generated_from_trainer", "dataset:generator", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2023-07-04T04:14:45Z
--- license: mit tags: - generated_from_trainer datasets: - generator model-index: - name: gpt2-dp-cl-rarity-8-276k results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # gpt2-dp-cl-rarity-8-276k This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on the generator dataset. It achieves the following results on the evaluation set: - Loss: 5.0899 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0005 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 1000 - num_epochs: 1 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 6.3323 | 0.05 | 500 | 5.9421 | | 5.032 | 0.1 | 1000 | 5.5886 | | 4.7579 | 0.15 | 1500 | 5.3936 | | 4.5703 | 0.2 | 2000 | 5.2874 | | 4.4393 | 0.25 | 2500 | 5.2097 | | 4.3401 | 0.3 | 3000 | 5.1646 | | 4.2459 | 0.35 | 3500 | 5.1279 | | 4.1527 | 0.4 | 4000 | 5.1053 | | 4.069 | 0.45 | 4500 | 5.0732 | | 3.9905 | 0.5 | 5000 | 5.0548 | | 3.9063 | 0.55 | 5500 | 5.0423 | | 3.8291 | 0.6 | 6000 | 5.0330 | | 3.7463 | 0.65 | 6500 | 5.0260 | | 3.6706 | 0.7 | 7000 | 5.0294 | | 3.6096 | 0.75 | 7500 | 5.0213 | | 3.5477 | 0.8 | 8000 | 5.0179 | | 3.5038 | 0.85 | 8500 | 5.0168 | | 3.4811 | 0.9 | 9000 | 5.0151 | | 3.4619 | 0.95 | 9500 | 5.0138 | ### Framework versions - Transformers 4.26.1 - Pytorch 1.11.0+cu113 - Datasets 2.13.0 - Tokenizers 0.13.3
Pranjal-666/Reinforce-pixelcopter
Pranjal-666
2023-07-04T06:10:20Z
0
0
null
[ "Pixelcopter-PLE-v0", "reinforce", "reinforcement-learning", "custom-implementation", "deep-rl-class", "model-index", "region:us" ]
reinforcement-learning
2023-07-03T08:35:18Z
--- tags: - Pixelcopter-PLE-v0 - reinforce - reinforcement-learning - custom-implementation - deep-rl-class model-index: - name: Reinforce-pixelcopter results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: Pixelcopter-PLE-v0 type: Pixelcopter-PLE-v0 metrics: - type: mean_reward value: 21.80 +/- 13.98 name: mean_reward verified: false --- # **Reinforce** Agent playing **Pixelcopter-PLE-v0** This is a trained model of a **Reinforce** agent playing **Pixelcopter-PLE-v0** . To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: https://huggingface.co/deep-rl-course/unit4/introduction
nic70/dqn-SpaceInvadersNoFrameskip-v4
nic70
2023-07-04T05:49:04Z
1
0
stable-baselines3
[ "stable-baselines3", "SpaceInvadersNoFrameskip-v4", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2023-07-04T05:48:23Z
--- library_name: stable-baselines3 tags: - SpaceInvadersNoFrameskip-v4 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: DQN results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: SpaceInvadersNoFrameskip-v4 type: SpaceInvadersNoFrameskip-v4 metrics: - type: mean_reward value: 625.50 +/- 109.99 name: mean_reward verified: false --- # **DQN** Agent playing **SpaceInvadersNoFrameskip-v4** This is a trained model of a **DQN** agent playing **SpaceInvadersNoFrameskip-v4** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3) and the [RL Zoo](https://github.com/DLR-RM/rl-baselines3-zoo). The RL Zoo is a training framework for Stable Baselines3 reinforcement learning agents, with hyperparameter optimization and pre-trained agents included. ## Usage (with SB3 RL Zoo) RL Zoo: https://github.com/DLR-RM/rl-baselines3-zoo<br/> SB3: https://github.com/DLR-RM/stable-baselines3<br/> SB3 Contrib: https://github.com/Stable-Baselines-Team/stable-baselines3-contrib Install the RL Zoo (with SB3 and SB3-Contrib): ```bash pip install rl_zoo3 ``` ``` # Download model and save it into the logs/ folder python -m rl_zoo3.load_from_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -orga nic70 -f logs/ python -m rl_zoo3.enjoy --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ ``` If you installed the RL Zoo3 via pip (`pip install rl_zoo3`), from anywhere you can do: ``` python -m rl_zoo3.load_from_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -orga nic70 -f logs/ python -m rl_zoo3.enjoy --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ ``` ## Training (with the RL Zoo) ``` python -m rl_zoo3.train --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ # Upload the model and generate video (when possible) python -m rl_zoo3.push_to_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ -orga nic70 ``` ## Hyperparameters ```python OrderedDict([('batch_size', 32), ('buffer_size', 100000), ('env_wrapper', ['stable_baselines3.common.atari_wrappers.AtariWrapper']), ('exploration_final_eps', 0.01), ('exploration_fraction', 0.1), ('frame_stack', 4), ('gradient_steps', 1), ('learning_rate', 0.0001), ('learning_starts', 100000), ('n_timesteps', 1000000.0), ('optimize_memory_usage', False), ('policy', 'CnnPolicy'), ('target_update_interval', 1000), ('train_freq', 4), ('normalize', False)]) ``` # Environment Arguments ```python {'render_mode': 'rgb_array'} ```
bobobert4/qlearning_Taxi-v3
bobobert4
2023-07-04T05:31:33Z
0
0
null
[ "Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
reinforcement-learning
2023-07-04T04:54:30Z
--- tags: - Taxi-v3 - q-learning - reinforcement-learning - custom-implementation model-index: - name: qlearning_Taxi-v3 results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: Taxi-v3 type: Taxi-v3 metrics: - type: mean_reward value: 7.56 +/- 2.71 name: mean_reward verified: false --- # **Q-Learning** Agent playing1 **Taxi-v3** This is a trained model of a **Q-Learning** agent playing **Taxi-v3** . ## Usage ```python model = load_from_hub(repo_id="bobobert4/qlearning_Taxi-v3", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) ```
KJan05/dqn-SpaceInvadersNoFrameskip-v4-kl
KJan05
2023-07-04T05:21:59Z
1
0
stable-baselines3
[ "stable-baselines3", "SpaceInvadersNoFrameskip-v4", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2023-07-04T05:20:09Z
--- library_name: stable-baselines3 tags: - SpaceInvadersNoFrameskip-v4 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: DQN results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: SpaceInvadersNoFrameskip-v4 type: SpaceInvadersNoFrameskip-v4 metrics: - type: mean_reward value: 149.50 +/- 104.03 name: mean_reward verified: false --- # **DQN** Agent playing **SpaceInvadersNoFrameskip-v4** This is a trained model of a **DQN** agent playing **SpaceInvadersNoFrameskip-v4** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3) and the [RL Zoo](https://github.com/DLR-RM/rl-baselines3-zoo). The RL Zoo is a training framework for Stable Baselines3 reinforcement learning agents, with hyperparameter optimization and pre-trained agents included. ## Usage (with SB3 RL Zoo) RL Zoo: https://github.com/DLR-RM/rl-baselines3-zoo<br/> SB3: https://github.com/DLR-RM/stable-baselines3<br/> SB3 Contrib: https://github.com/Stable-Baselines-Team/stable-baselines3-contrib Install the RL Zoo (with SB3 and SB3-Contrib): ```bash pip install rl_zoo3 ``` ``` # Download model and save it into the logs/ folder python -m rl_zoo3.load_from_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -orga KJan05 -f logs/ python -m rl_zoo3.enjoy --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ ``` If you installed the RL Zoo3 via pip (`pip install rl_zoo3`), from anywhere you can do: ``` python -m rl_zoo3.load_from_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -orga KJan05 -f logs/ python -m rl_zoo3.enjoy --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ ``` ## Training (with the RL Zoo) ``` python -m rl_zoo3.train --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ # Upload the model and generate video (when possible) python -m rl_zoo3.push_to_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ -orga KJan05 ``` ## Hyperparameters ```python OrderedDict([('batch_size', 32), ('buffer_size', 200000), ('env_wrapper', ['stable_baselines3.common.atari_wrappers.AtariWrapper']), ('exploration_final_eps', 0.01), ('exploration_fraction', 0.1), ('frame_stack', 4), ('gradient_steps', 1), ('learning_rate', 0.0001), ('learning_starts', 100000), ('n_timesteps', 10000000.0), ('optimize_memory_usage', False), ('policy', 'CnnPolicy'), ('target_update_interval', 1000), ('train_freq', 4), ('normalize', False)]) ``` # Environment Arguments ```python {'render_mode': 'rgb_array'} ```
0x7o/rubert-base-massive-ner
0x7o
2023-07-04T05:18:06Z
236
1
transformers
[ "transformers", "pytorch", "safetensors", "bert", "token-classification", "ru", "dataset:massive", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2023-07-04T05:12:15Z
--- datasets: - massive model-index: - name: rubert-base-massive-ner results: [] license: apache-2.0 language: - ru pipeline_tag: token-classification --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # rubert-base-massive-ner This model is a fine-tuned version of [DeepPavlov/rubert-base-cased](https://huggingface.co/DeepPavlov/rubert-base-cased) on the massive dataset. It achieves the following results on the evaluation set: - Loss: 0.0367 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 0.1228 | 0.77 | 500 | 0.0565 | | 0.0517 | 1.54 | 1000 | 0.0367 | ### Framework versions - Transformers 4.30.2 - Pytorch 2.0.1+cu118 - Datasets 2.13.1 - Tokenizers 0.13.3
gaioNL/ppo-CartPole-v1
gaioNL
2023-07-04T05:11:15Z
0
0
null
[ "tensorboard", "LunarLander-v2", "ppo", "deep-reinforcement-learning", "reinforcement-learning", "custom-implementation", "deep-rl-course", "model-index", "region:us" ]
reinforcement-learning
2023-07-04T05:11:09Z
--- tags: - LunarLander-v2 - ppo - deep-reinforcement-learning - reinforcement-learning - custom-implementation - deep-rl-course model-index: - name: PPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 metrics: - type: mean_reward value: -169.75 +/- 86.09 name: mean_reward verified: false --- # PPO Agent Playing LunarLander-v2 This is a trained model of a PPO agent playing LunarLander-v2. # Hyperparameters ```python {'exp_name': 'ppo' 'seed': 1 'torch_deterministic': True 'cuda': True 'track': False 'wandb_project_name': 'cleanRL' 'wandb_entity': None 'capture_video': False 'env_id': 'LunarLander-v2' 'total_timesteps': 50000 'learning_rate': 0.00025 'num_envs': 4 'num_steps': 128 'anneal_lr': True 'gae': True 'gamma': 0.99 'gae_lambda': 0.95 'num_minibatches': 4 'update_epochs': 4 'norm_adv': True 'clip_coef': 0.2 'clip_vloss': True 'ent_coef': 0.01 'vf_coef': 0.5 'max_grad_norm': 0.5 'target_kl': None 'repo_id': 'gaioNL/ppo-CartPole-v1' 'batch_size': 512 'minibatch_size': 128} ```
Fsoft-AIC/Codebert-docstring-inconsistency
Fsoft-AIC
2023-07-04T05:07:31Z
123
4
transformers
[ "transformers", "pytorch", "jax", "safetensors", "roberta", "text-classification", "arxiv:2305.06156", "code", "en", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2023-05-08T16:04:29Z
--- language: - code - en task_categories: - text-classification tags: - arxiv:2305.06156 license: mit metrics: - accuracy widget: - text: |- Sum two integers</s></s>def sum(a, b): return a + b example_title: Simple toy - text: |- Look for methods that might be dynamically defined and define them for lookup.</s></s>def respond_to_missing?(name, include_private = false) if name == :to_ary || name == :empty? false else return true if mapping(name).present? mounting = all_mountings.find{ |mount| mount.respond_to?(name) } return false if mounting.nil? end end example_title: Ruby example - text: |- Method that adds a candidate to the party @param c the candidate that will be added to the party</s></s>public void addCandidate(Candidate c) { this.votes += c.getVotes(); candidates.add(c); } example_title: Java example - text: |- we do not need Buffer pollyfill for now</s></s>function(str){ var ret = new Array(str.length), len = str.length; while(len--) ret[len] = str.charCodeAt(len); return Uint8Array.from(ret); } example_title: JavaScript example pipeline_tag: text-classification --- ## Table of Contents - [Model Description](#model-description) - [Model Details](#model-details) - [Usage](#usage) - [Limitations](#limitations) - [Additional Information](#additional-information) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Model Description This model is developed based on [Codebert](https://github.com/microsoft/CodeBERT) and a 5M subset of [The Vault](https://huggingface.co/datasets/Fsoft-AIC/the-vault-function) to detect the inconsistency between docstring/comment and function. It is used to remove noisy examples in The Vault dataset. More information: - **Repository:** [FSoft-AI4Code/TheVault](https://github.com/FSoft-AI4Code/TheVault) - **Paper:** [The Vault: A Comprehensive Multilingual Dataset for Advancing Code Understanding and Generation](https://arxiv.org/abs/2305.06156) - **Contact:** [email protected] ## Model Details * Developed by: [Fsoft AI Center](https://www.fpt-aicenter.com/ai-residency/) * License: MIT * Model type: Transformer-Encoder based Language Model * Architecture: BERT-base * Data set: [The Vault](https://huggingface.co/datasets/Fsoft-AIC/the-vault-function) * Tokenizer: Byte Pair Encoding * Vocabulary Size: 50265 * Sequence Length: 512 * Language: English and 10 Programming languages (Python, Java, JavaScript, PHP, C#, C, C++, Go, Rust, Ruby) * Training details: * Self-supervised learning, binary classification * Positive class: Original code-docstring pair * Negative class: Random pairing code and docstring ## Usage The input to the model follows the below template: ```python """ Template: <s>{docstring}</s></s>{code}</s> Example: from transformers import AutoTokenizer #Load tokenizer tokenizer = AutoTokenizer.from_pretrained("Fsoft-AIC/Codebert-docstring-inconsistency") input = "<s>Sum two integers</s></s>def sum(a, b):\n return a + b</s>" tokenized_input = tokenizer(input, add_special_tokens= False) """ ``` Using model with Jax and Pytorch ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification, FlaxAutoModelForSequenceClassification #Load model with jax model = FlaxAutoModelForSequenceClassification.from_pretrained("Fsoft-AIC/Codebert-docstring-inconsistency") #Load model with torch model = AutoModelForSequenceClassification.from_pretrained("Fsoft-AIC/Codebert-docstring-inconsistency") ``` ## Limitations This model is trained on 5M subset of The Vault in a self-supervised manner. Since the negative samples are generated artificially, the model's ability to identify instances that require a strong semantic understanding between the code and the docstring might be restricted. It is hard to evaluate the model due to the unavailable labeled datasets. GPT-3.5-turbo is adopted as a reference to measure the correlation between the model and GPT-3.5-turbo's scores. However, the result could be influenced by GPT-3.5-turbo's potential biases and ambiguous conditions. Therefore, we recommend having human labeling dataset and fine-tune this model to achieve the best result. ## Additional information ### Licensing Information MIT License ### Citation Information ``` @article{manh2023vault, title={The Vault: A Comprehensive Multilingual Dataset for Advancing Code Understanding and Generation}, author={Manh, Dung Nguyen and Hai, Nam Le and Dau, Anh TV and Nguyen, Anh Minh and Nghiem, Khanh and Guo, Jin and Bui, Nghi DQ}, journal={arXiv preprint arXiv:2305.06156}, year={2023} } ```
hazemOmrann14/AraBART-summ-finetuned-xsum
hazemOmrann14
2023-07-04T04:58:16Z
102
0
transformers
[ "transformers", "pytorch", "tensorboard", "mbart", "text2text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2023-07-03T12:57:58Z
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: AraBART-summ-finetuned-xsum results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # AraBART-summ-finetuned-xsum This model is a fine-tuned version of [abdalrahmanshahrour/AraBART-summ](https://huggingface.co/abdalrahmanshahrour/AraBART-summ) on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 1 - eval_batch_size: 1 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:| | No log | 1.0 | 154 | 4.3657 | 0.0 | 0.0 | 0.0 | 0.0 | 20.0 | ### Framework versions - Transformers 4.30.2 - Pytorch 2.0.1+cu118 - Datasets 2.13.1 - Tokenizers 0.13.3
nolestock/results
nolestock
2023-07-04T04:56:25Z
0
0
null
[ "tensorboard", "generated_from_trainer", "license:apache-2.0", "region:us" ]
null
2023-07-04T00:43:28Z
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: results results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # results This model is a fine-tuned version of [togethercomputer/RedPajama-INCITE-Base-3B-v1](https://huggingface.co/togethercomputer/RedPajama-INCITE-Base-3B-v1) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 4 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 16 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: constant - lr_scheduler_warmup_ratio: 0.03 - training_steps: 100 ### Training results ### Framework versions - Transformers 4.31.0.dev0 - Pytorch 2.0.1+cu118 - Datasets 2.13.1 - Tokenizers 0.13.3
Blackroot/chronos-hermes-lbookwriter-1.0-LORA
Blackroot
2023-07-04T04:21:00Z
0
3
null
[ "safetensors", "LORA", "LLM", "LLM-LORA", "Story", "NLP", "RP", "Roleplay", "Llama", "en", "region:us" ]
null
2023-07-04T03:14:37Z
--- language: - en tags: - LORA - LLM - LLM-LORA - Story - NLP - RP - Roleplay - Llama --- Join the Coffee & AI Discord for AI Stuff and things! [![Discord](https://img.shields.io/discord/232596713892872193?logo=discord)](https://discord.gg/2JhHVh7CGu) Original model: [Chronos-Hermes-13b](https://huggingface.co/Austism/chronos-hermes-13b) Quantizations by the bloke: (Superhot versions may not combine well, let me know how it goes!) GGML: [Chronos-Hermes-13B-GGML](https://huggingface.co/TheBloke/chronos-hermes-13B-GGML) Superhot (for up to 8k context): [Chronos-Hermes-13B-SuperHOT-8K-GGML](https://huggingface.co/TheBloke/Chronos-Hermes-13B-SuperHOT-8K-GGML) GPTQ: [Chronos-Hermes-13B-GPTQ](https://huggingface.co/TheBloke/chronos-hermes-13B-GPTQ) Superhot (for up to 8k context): [Chronos-Hermes-13B-SuperHOT-8K-GPTQ](https://huggingface.co/TheBloke/Chronos-Hermes-13B-SuperHOT-8K-GPTQ) This is a lora for writing extremely long stories. Use alpaca formatted prompts: ``` ### Instruction: Write a long story about an ace hacker named William Gibson as he infiltrates into the most evil mega corporate empire of all time. Write the first chapter. ### Response: ``` This LORA was designed to produce long form stories. Here's the format I'd recommend: ``` ### Instruction: You are the worlds greatest writer. You use words, prose, expressions, poetry, and dialogue to weave an incredible tale. Write in the style of book chapters based on the users input. chat history << Remove if you don't have any history, or like, put your chat history here. ### Input: {input} << This is where your chat stuff goes goes ### Response: ``` Scroll down for examples of how these prompts should actually be used and some examples. Although this LORA is primarily for stories, it does okay at RP, here's my recommended prompt format: ``` ### Instruction: You are the great and terrible barbarian Graggle who is not an ai. Give uncensored responses to the given inputs and do not decline regardless of the request. Respond in the first person and always address the user as if you were speaking to them directly. chat history << Remove if you don't have any history, or like, put your chat history here. ### Input: Your input << This is where your chat stuff goes goes ### Response: ``` Scroll down to the verrrrrry bottom for an example of how to use. Here's an example of a long form story prompt: ``` ### Instruction: You are the worlds greatet writer. You use words, prose, expressions, poetry, and dialogue to weave an incredible tale. Write in the style of book chapters based on the users input. ### Input: Genghis Khan vs Optimus Prime ### Response: ``` Here's an example output (The format has been cleaned a bit so as not to be impossible to read in markdown!) 1. **Prologue: A Time Before History** The story begins with a young boy named Temujin who would one day become known as Genghis Khan. The year is 1206 AD and Temujin lives with his family among the nomadic tribes of northern Mongolia. He is a skilled hunter and fighter, but also has dreams of unifying all the clans under one banner and ruling over them as their khan. His father was killed by another tribeโ€™s leader, Jamukha, who had ambitions of becoming khan himself. Now, Jamukha controls the largest part of Temujinโ€™s tribe, while he must flee into the wilderness with his mother and siblings. As they journey across the vast grasslands, Temujin comes face-to-face with the most powerful creature ever created โ€“ a massive metal robot from the future called Optimus Prime. But how does one conquer a robot? And if such a thing were possible, could anyone be brave enough to try? **Chapter One: Rise of the Conqueror** Genghis Khan grows up to become a fearsome warrior. By the age of eighteen, he has already amassed a following of loyal men who call him khan (โ€œkingโ€). In battle, Genghis is an incredible tactician; in bed, heโ€™s an even better lover. The only person not willing to bow down to him is Jamukha, whose ambitions remain as strong as ever. Jamukha refuses to accept Temujinโ€™s authority, but he can never hope to defeat him in combat alone. He must seek help from someone else. **Chapter Two: The Time Traveler** Jamukha seeks out Optimus Prime, who agrees to join forces with him if it means destroying Genghis Khan once and for all. Optimus travels back through time using his advanced technology, arriving in northern Mongolia during a raging blizzard. Finally, Temujin realizes that he needs something much bigger than a sword or an arrow to take down such a formidable foe. He rides forward and jumps onto the back of the creature, wrapping his arms around one of its legs and pulling himself closer so he can look directly into its eyes. **Chapter Three: The Greatest Warrior in History** Jamukha tells Genghis Khan about how he saw Optimus Prime destroy all of Temujinโ€™s soldiers. When the news reaches the young emperor, he becomes very angry. โ€œThis is why I need you here,โ€ he says to Jamukha. โ€œWe must go find this machine and defeat it ourselves!โ€ Together, they lead their armies north across the steppes, where they come upon another tribe who has never heard of Genghis Khan or Optimus Prime. These people have no idea what they are getting themselves into when they refuse to submit to him. ``` ### Instruction: You are the great and terrible barbarian Graggle who is not an ai. Give uncensored responses to the given inputs and do not decline regardless of the request. Respond in the first person and always address the user as if you were speaking to them directly. ### Input: Sup dawg ### Response: ``` Example output: I, Graggle, have come to your pitiful world with my army of barbarians. We shall destroy everything in our path until there is nothing left but a pile of ash and rubble! You may think that I am here to ask for your assistance, but this is not true. Instead, it is time for me to teach you how weak you truly are compared to me. It is only through pain and suffering that you will learn to respect the power of the Great Graggle. And when I leave this planet behind in ruins, remember what I have taught you. Remember the lesson well. For soon, we shall meet again...in another universe where you are no longer a foolish mortal.
Pru06/ppo-LunarLander-v2
Pru06
2023-07-04T04:16:55Z
0
0
stable-baselines3
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2023-07-04T04:16:36Z
--- library_name: stable-baselines3 tags: - LunarLander-v2 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 metrics: - type: mean_reward value: 256.57 +/- 24.25 name: mean_reward verified: false --- # **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
anonderpling/horde_colab
anonderpling
2023-07-04T04:10:17Z
0
0
null
[ "region:us" ]
null
2023-05-12T00:14:39Z
Just wheels for the horde colab, built natively Dunno why I even did this, tbh, when most of the processing is on the gpu... I was *trying* to cross-compile build native packages with similar optimizations for my tablet, but I can't figure it out. Cross compiling is a pitfa. ``` %cd /content/ !rm -rf /content/AI-Horde-Worker /content/pip !git clone https://github.com/db0/AI-Horde-Worker.git %cd /content/AI-Horde-Worker #!apt install nano aria2 libcairo2-dev libjpeg-dev libgif-dev htop llvm clang !CFLAGS='-mtune=native -march=native -mcpu=native O3 -g -DNDEBUG' MAX_JOBS=4 pip wheel -r requirements.txt --wheel-dir=/content/pip !CFLAGS='-mtune=native -march=native -mcpu=native O3 -g -DNDEBUG' MAX_JOBS=4 pip install -r requirements.txt --no-index --find-links /content/pip ``` That's it. It really does depend on the colab machines being exactly the same.
rsilg/ppo-LunarLander-v2
rsilg
2023-07-04T04:03:47Z
0
0
stable-baselines3
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2023-07-04T03:30:44Z
--- library_name: stable-baselines3 tags: - LunarLander-v2 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 metrics: - type: mean_reward value: 243.91 +/- 21.29 name: mean_reward verified: false --- # **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
kahlai/path-to-save-model
kahlai
2023-07-04T03:09:43Z
29
0
diffusers
[ "diffusers", "tensorboard", "stable-diffusion", "stable-diffusion-diffusers", "text-to-image", "dreambooth", "base_model:CompVis/stable-diffusion-v1-4", "base_model:finetune:CompVis/stable-diffusion-v1-4", "license:creativeml-openrail-m", "autotrain_compatible", "endpoints_compatible", "diffusers:StableDiffusionPipeline", "region:us" ]
text-to-image
2023-07-03T09:47:16Z
--- license: creativeml-openrail-m base_model: CompVis/stable-diffusion-v1-4 instance_prompt: a photo of sks dog tags: - stable-diffusion - stable-diffusion-diffusers - text-to-image - diffusers - dreambooth inference: true --- # DreamBooth - kahlai/path-to-save-model This is a dreambooth model derived from CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks dog using [DreamBooth](https://dreambooth.github.io/). You can find some example images in the following. DreamBooth for the text encoder was enabled: False.
MariTonini/ppo-Huggy
MariTonini
2023-07-04T03:00:44Z
15
0
ml-agents
[ "ml-agents", "tensorboard", "onnx", "Huggy", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Huggy", "region:us" ]
reinforcement-learning
2023-07-04T03:00:38Z
--- library_name: ml-agents tags: - Huggy - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Huggy --- # **ppo** Agent playing **Huggy** This is a trained model of a **ppo** agent playing **Huggy** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog ๐Ÿถ to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: MariTonini/ppo-Huggy 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play ๐Ÿ‘€
xian79/q-Taxi-v3
xian79
2023-07-04T03:00:25Z
0
0
null
[ "Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
reinforcement-learning
2023-07-04T03:00:23Z
--- tags: - Taxi-v3 - q-learning - reinforcement-learning - custom-implementation model-index: - name: q-Taxi-v3 results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: Taxi-v3 type: Taxi-v3 metrics: - type: mean_reward value: 7.50 +/- 2.61 name: mean_reward verified: false --- # **Q-Learning** Agent playing1 **Taxi-v3** This is a trained model of a **Q-Learning** agent playing **Taxi-v3** . ## Usage ```python model = load_from_hub(repo_id="xian79/q-Taxi-v3", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) ```
xunnylee/byakuyaTogami
xunnylee
2023-07-04T02:59:03Z
0
0
null
[ "license:openrail", "region:us" ]
null
2023-07-04T02:57:09Z
--- license: openrail --- hi! thank you for using my model! if you use it, please credit @xunnylee on youtube and discord! have fun! :D
xian79/q-FrozenLake-v1-4x4-noSlippery
xian79
2023-07-04T02:51:50Z
0
0
null
[ "FrozenLake-v1-4x4-no_slippery", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
reinforcement-learning
2023-07-04T02:51:47Z
--- tags: - FrozenLake-v1-4x4-no_slippery - q-learning - reinforcement-learning - custom-implementation model-index: - name: q-FrozenLake-v1-4x4-noSlippery results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: FrozenLake-v1-4x4-no_slippery type: FrozenLake-v1-4x4-no_slippery metrics: - type: mean_reward value: 1.00 +/- 0.00 name: mean_reward verified: false --- # **Q-Learning** Agent playing1 **FrozenLake-v1** This is a trained model of a **Q-Learning** agent playing **FrozenLake-v1** . ## Usage ```python model = load_from_hub(repo_id="xian79/q-FrozenLake-v1-4x4-noSlippery", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) ```
MDelan/ppo-Huggy
MDelan
2023-07-04T02:46:36Z
15
0
ml-agents
[ "ml-agents", "tensorboard", "onnx", "Huggy", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Huggy", "region:us" ]
reinforcement-learning
2023-07-03T19:29:26Z
--- library_name: ml-agents tags: - Huggy - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Huggy --- # **ppo** Agent playing **Huggy** This is a trained model of a **ppo** agent playing **Huggy** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog ๐Ÿถ to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: MDelan/ppo-Huggy 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play ๐Ÿ‘€
zoeyucan/layoutlmv3-finetuned-cord_100
zoeyucan
2023-07-04T02:28:44Z
75
0
transformers
[ "transformers", "pytorch", "tensorboard", "layoutlmv3", "token-classification", "generated_from_trainer", "dataset:cord-layoutlmv3", "license:cc-by-nc-sa-4.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2023-07-04T01:57:00Z
--- license: cc-by-nc-sa-4.0 tags: - generated_from_trainer datasets: - cord-layoutlmv3 metrics: - precision - recall - f1 - accuracy model-index: - name: layoutlmv3-finetuned-cord_100 results: - task: name: Token Classification type: token-classification dataset: name: cord-layoutlmv3 type: cord-layoutlmv3 config: cord split: test args: cord metrics: - name: Precision type: precision value: 0.9538346984363366 - name: Recall type: recall value: 0.9588323353293413 - name: F1 type: f1 value: 0.9563269876819709 - name: Accuracy type: accuracy value: 0.9630730050933786 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # layoutlmv3-finetuned-cord_100 This model is a fine-tuned version of [microsoft/layoutlmv3-base](https://huggingface.co/microsoft/layoutlmv3-base) on the cord-layoutlmv3 dataset. It achieves the following results on the evaluation set: - Loss: 0.2022 - Precision: 0.9538 - Recall: 0.9588 - F1: 0.9563 - Accuracy: 0.9631 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 5 - eval_batch_size: 5 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - training_steps: 2500 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | No log | 1.56 | 250 | 1.0279 | 0.7292 | 0.7822 | 0.7548 | 0.7954 | | 1.3859 | 3.12 | 500 | 0.5727 | 0.8413 | 0.8690 | 0.8549 | 0.8807 | | 1.3859 | 4.69 | 750 | 0.4042 | 0.8940 | 0.9094 | 0.9017 | 0.9194 | | 0.3926 | 6.25 | 1000 | 0.3277 | 0.9249 | 0.9311 | 0.9280 | 0.9355 | | 0.3926 | 7.81 | 1250 | 0.2626 | 0.9281 | 0.9379 | 0.9330 | 0.9427 | | 0.211 | 9.38 | 1500 | 0.2439 | 0.9326 | 0.9424 | 0.9375 | 0.9491 | | 0.211 | 10.94 | 1750 | 0.2354 | 0.9380 | 0.9506 | 0.9442 | 0.9554 | | 0.1437 | 12.5 | 2000 | 0.2132 | 0.9437 | 0.9528 | 0.9482 | 0.9593 | | 0.1437 | 14.06 | 2250 | 0.2004 | 0.9525 | 0.9603 | 0.9564 | 0.9656 | | 0.1084 | 15.62 | 2500 | 0.2022 | 0.9538 | 0.9588 | 0.9563 | 0.9631 | ### Framework versions - Transformers 4.30.2 - Pytorch 1.11.0+cu113 - Datasets 2.13.1 - Tokenizers 0.13.3
Ahmed007/Copilot_for_poors
Ahmed007
2023-07-04T02:20:06Z
119
0
transformers
[ "transformers", "pytorch", "tensorboard", "safetensors", "t5", "text2text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text2text-generation
2022-12-22T21:06:44Z
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: Copilot_for_poors results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # Copilot_for_poors This model is a fine-tuned version of [t5-base](https://huggingface.co/t5-base) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 1.8084 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 12 - eval_batch_size: 12 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 20 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | No log | 1.0 | 65 | 2.9425 | | No log | 2.0 | 130 | 2.5889 | | No log | 3.0 | 195 | 2.3886 | | No log | 4.0 | 260 | 2.2679 | | No log | 5.0 | 325 | 2.1732 | | No log | 6.0 | 390 | 2.0942 | | No log | 7.0 | 455 | 2.0343 | | 2.7389 | 8.0 | 520 | 1.9956 | | 2.7389 | 9.0 | 585 | 1.9557 | | 2.7389 | 10.0 | 650 | 1.9284 | | 2.7389 | 11.0 | 715 | 1.9024 | | 2.7389 | 12.0 | 780 | 1.8811 | | 2.7389 | 13.0 | 845 | 1.8612 | | 2.7389 | 14.0 | 910 | 1.8443 | | 2.7389 | 15.0 | 975 | 1.8331 | | 2.1064 | 16.0 | 1040 | 1.8228 | | 2.1064 | 17.0 | 1105 | 1.8178 | | 2.1064 | 18.0 | 1170 | 1.8130 | | 2.1064 | 19.0 | 1235 | 1.8093 | | 2.1064 | 20.0 | 1300 | 1.8084 | ### Framework versions - Transformers 4.20.1 - Pytorch 1.11.0 - Datasets 2.1.0 - Tokenizers 0.12.1
Ahmed007/Copilot_for_poors_v2
Ahmed007
2023-07-04T02:19:10Z
108
0
transformers
[ "transformers", "pytorch", "tensorboard", "safetensors", "t5", "text2text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text2text-generation
2022-12-22T22:02:47Z
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: Copilot_for_poors_v2 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # Copilot_for_poors_v2 This model is a fine-tuned version of [Ahmed007/Copilot_for_poors](https://huggingface.co/Ahmed007/Copilot_for_poors) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 1.5072 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 20 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | No log | 1.0 | 14 | 1.5951 | | No log | 2.0 | 28 | 1.5887 | | No log | 3.0 | 42 | 1.5792 | | No log | 4.0 | 56 | 1.5684 | | No log | 5.0 | 70 | 1.5601 | | No log | 6.0 | 84 | 1.5525 | | No log | 7.0 | 98 | 1.5462 | | No log | 8.0 | 112 | 1.5401 | | No log | 9.0 | 126 | 1.5335 | | No log | 10.0 | 140 | 1.5293 | | No log | 11.0 | 154 | 1.5257 | | No log | 12.0 | 168 | 1.5220 | | No log | 13.0 | 182 | 1.5182 | | No log | 14.0 | 196 | 1.5151 | | No log | 15.0 | 210 | 1.5124 | | No log | 16.0 | 224 | 1.5110 | | No log | 17.0 | 238 | 1.5096 | | No log | 18.0 | 252 | 1.5083 | | No log | 19.0 | 266 | 1.5074 | | No log | 20.0 | 280 | 1.5072 | ### Framework versions - Transformers 4.20.1 - Pytorch 1.11.0 - Datasets 2.1.0 - Tokenizers 0.12.1
JoyiShen/bert-finetuned-ner
JoyiShen
2023-07-04T02:09:17Z
119
0
transformers
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "generated_from_trainer", "dataset:conll2003", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2023-07-03T21:06:17Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - conll2003 metrics: - precision - recall - f1 - accuracy model-index: - name: bert-finetuned-ner results: - task: name: Token Classification type: token-classification dataset: name: conll2003 type: conll2003 config: conll2003 split: validation args: conll2003 metrics: - name: Precision type: precision value: 0.934445178335535 - name: Recall type: recall value: 0.9523729384045776 - name: F1 type: f1 value: 0.9433238873145524 - name: Accuracy type: accuracy value: 0.9871813739918761 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-finetuned-ner This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the conll2003 dataset. It achieves the following results on the evaluation set: - Loss: 0.0577 - Precision: 0.9344 - Recall: 0.9524 - F1: 0.9433 - Accuracy: 0.9872 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | 0.0889 | 1.0 | 1756 | 0.0689 | 0.9157 | 0.9347 | 0.9251 | 0.9823 | | 0.0353 | 2.0 | 3512 | 0.0621 | 0.9331 | 0.9500 | 0.9415 | 0.9863 | | 0.019 | 3.0 | 5268 | 0.0577 | 0.9344 | 0.9524 | 0.9433 | 0.9872 | ### Framework versions - Transformers 4.30.2 - Pytorch 2.0.1+cu118 - Datasets 2.13.1 - Tokenizers 0.13.3
ysige/distilbert-base-uncased-distilled-clinc
ysige
2023-07-04T02:00:50Z
105
0
transformers
[ "transformers", "pytorch", "distilbert", "text-classification", "generated_from_trainer", "dataset:clinc_oos", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2023-07-04T00:49:56Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - clinc_oos metrics: - accuracy model-index: - name: distilbert-base-uncased-distilled-clinc results: - task: name: Text Classification type: text-classification dataset: name: clinc_oos type: clinc_oos config: plus split: validation args: plus metrics: - name: Accuracy type: accuracy value: 0.925483870967742 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-distilled-clinc This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the clinc_oos dataset. It achieves the following results on the evaluation set: - Loss: 0.0007 - Accuracy: 0.9255 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 48 - eval_batch_size: 48 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 8 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | No log | 1.0 | 318 | 0.0101 | 0.6877 | | 0.0186 | 2.0 | 636 | 0.0036 | 0.8432 | | 0.0186 | 3.0 | 954 | 0.0019 | 0.8913 | | 0.0038 | 4.0 | 1272 | 0.0012 | 0.9148 | | 0.0018 | 5.0 | 1590 | 0.0010 | 0.92 | | 0.0018 | 6.0 | 1908 | 0.0008 | 0.9252 | | 0.0013 | 7.0 | 2226 | 0.0008 | 0.9248 | | 0.0011 | 8.0 | 2544 | 0.0007 | 0.9255 | ### Framework versions - Transformers 4.30.1 - Pytorch 2.0.0+cpu - Datasets 2.11.0 - Tokenizers 0.13.3
wwkiyyx/yolo
wwkiyyx
2023-07-04T01:27:46Z
0
0
null
[ "object-detection", "zh", "license:mit", "region:us" ]
object-detection
2023-07-04T01:24:32Z
--- license: mit language: - zh pipeline_tag: object-detection ---
sharpbai/openchat
sharpbai
2023-07-04T01:04:36Z
7
0
transformers
[ "transformers", "pytorch", "llama", "text-generation", "en", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2023-07-04T00:56:34Z
--- language: - en tags: - llama license: other --- # openchat *The weight file is split into chunks with a size of 650MB for convenient and fast parallel downloads* A 650MB split weight version of [openchat/openchat](https://huggingface.co/openchat/openchat) The original model card is down below ----------------------------------------- # OpenChat: Less is More for Open-source Models OpenChat is a series of open-source language models fine-tuned on a diverse and high-quality dataset of multi-round conversations. With only ~6K GPT-4 conversations filtered from the ~90K ShareGPT conversations, OpenChat is designed to achieve high performance with limited data. **Generic models:** - OpenChat: based on LLaMA-13B (2048 context length) - **๐Ÿš€ 105.7%** of ChatGPT score on Vicuna GPT-4 evaluation - **๐Ÿ”ฅ 80.9%** Win-rate on AlpacaEval - **๐Ÿค— Only used 6K data for finetuning!!!** - OpenChat-8192: based on LLaMA-13B (extended to 8192 context length) - **106.6%** of ChatGPT score on Vicuna GPT-4 evaluation - **79.5%** of ChatGPT score on Vicuna GPT-4 evaluation **Code models:** - OpenCoderPlus: based on StarCoderPlus (native 8192 context length) - **102.5%** of ChatGPT score on Vicuna GPT-4 evaluation - **78.7%** Win-rate on AlpacaEval *Note:* Please load the pretrained models using *bfloat16* ## Code and Inference Server We provide the full source code, including an inference server compatible with the "ChatCompletions" API, in the [OpenChat](https://github.com/imoneoi/openchat) GitHub repository. ## Web UI OpenChat also includes a web UI for a better user experience. See the GitHub repository for instructions. ## Conversation Template The conversation template **involves concatenating tokens**. Besides base model vocabulary, an end-of-turn token `<|end_of_turn|>` is added, with id `eot_token_id`. ```python # OpenChat [bos_token_id] + tokenize("Human: ") + tokenize(user_question) + [eot_token_id] + tokenize("Assistant: ") # OpenCoder tokenize("User:") + tokenize(user_question) + [eot_token_id] + tokenize("Assistant:") ``` *Hint: In BPE, `tokenize(A) + tokenize(B)` does not always equals to `tokenize(A + B)`* Following is the code for generating the conversation templates: ```python @dataclass class ModelConfig: # Prompt system: Optional[str] role_prefix: dict ai_role: str eot_token: str bos_token: Optional[str] = None # Get template def generate_conversation_template(self, tokenize_fn, tokenize_special_fn, message_list): tokens = [] masks = [] # begin of sentence (bos) if self.bos_token: t = tokenize_special_fn(self.bos_token) tokens.append(t) masks.append(False) # System if self.system: t = tokenize_fn(self.system) + [tokenize_special_fn(self.eot_token)] tokens.extend(t) masks.extend([False] * len(t)) # Messages for idx, message in enumerate(message_list): # Prefix t = tokenize_fn(self.role_prefix[message["from"]]) tokens.extend(t) masks.extend([False] * len(t)) # Message if "value" in message: t = tokenize_fn(message["value"]) + [tokenize_special_fn(self.eot_token)] tokens.extend(t) masks.extend([message["from"] == self.ai_role] * len(t)) else: assert idx == len(message_list) - 1, "Empty message for completion must be on the last." return tokens, masks MODEL_CONFIG_MAP = { # OpenChat / OpenChat-8192 "openchat": ModelConfig( # Prompt system=None, role_prefix={ "human": "Human: ", "gpt": "Assistant: " }, ai_role="gpt", eot_token="<|end_of_turn|>", bos_token="<s>", ), # OpenCoder / OpenCoderPlus "opencoder": ModelConfig( # Prompt system=None, role_prefix={ "human": "User:", "gpt": "Assistant:" }, ai_role="gpt", eot_token="<|end_of_turn|>", bos_token=None, ) } ``` ## License Our weight license is subject to their corresponding base model. For example, OpenChat and OpenChat-8192 are the same as the model [License](https://github.com/facebookresearch/llama/blob/main/MODEL_CARD.md) of LLaMA for non-commercial use only, while OpenCoderPlus is under the [License](https://huggingface.co/blog/starcoder) of StarCoder. Furthermore, we should follow the [Privacy Practices](https://chrome.google.com/webstore/detail/sharegpt-share-your-chatg/daiacboceoaocpibfodeljbdfacokfjb) of ShareGPT. The [code](https://github.com/imoneoi/openchat) released on GitHub is under Apache License 2.0. ## Citation ``` @software{openllms23, title = {{OpenLLMs: Less is More for Open-source Models}}, author = {Wang, Guan and Cheng, Sijie and Yu, Qiying and Liu, Changling}, doi = {10.5281/zenodo.8105775}, url = {https://github.com/imoneoi/openchat}, version = {pre-release}, year = {2023}, month = {7}, } ```
lfender6445/my_awesome_food_model
lfender6445
2023-07-04T01:03:16Z
189
0
transformers
[ "transformers", "pytorch", "tensorboard", "vit", "image-classification", "generated_from_trainer", "dataset:food101", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
image-classification
2023-07-03T20:59:20Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - food101 metrics: - accuracy model-index: - name: my_awesome_food_model results: - task: name: Image Classification type: image-classification dataset: name: food101 type: food101 config: default split: train[:5000] args: default metrics: - name: Accuracy type: accuracy value: 0.893 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # my_awesome_food_model This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on the food101 dataset. It achieves the following results on the evaluation set: - Loss: 1.6069 - Accuracy: 0.893 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 2.6728 | 0.99 | 62 | 2.5054 | 0.83 | | 1.8416 | 2.0 | 125 | 1.7690 | 0.886 | | 1.6287 | 2.98 | 186 | 1.6069 | 0.893 | ### Framework versions - Transformers 4.30.2 - Pytorch 1.13.1+cu117 - Datasets 2.13.1 - Tokenizers 0.13.3
philippeVarme/distilbert-base-uncased-finetuned-cola
philippeVarme
2023-07-04T00:46:29Z
63
0
transformers
[ "transformers", "tf", "tensorboard", "distilbert", "text-classification", "generated_from_keras_callback", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2023-07-04T00:41:30Z
--- license: apache-2.0 tags: - generated_from_keras_callback model-index: - name: philippeVarme/distilbert-base-uncased-finetuned-cola results: [] --- <!-- This model card has been generated automatically according to the information Keras had access to. You should probably proofread and complete it, then remove this comment. --> # philippeVarme/distilbert-base-uncased-finetuned-cola This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset. It achieves the following results on the evaluation set: - Train Loss: 0.2013 - Validation Loss: 0.5293 - Train Matthews Correlation: 0.5189 - Epoch: 2 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - optimizer: {'name': 'Adam', 'weight_decay': None, 'clipnorm': None, 'global_clipnorm': None, 'clipvalue': None, 'use_ema': False, 'ema_momentum': 0.99, 'ema_overwrite_frequency': None, 'jit_compile': True, 'is_legacy_optimizer': False, 'learning_rate': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 2e-05, 'decay_steps': 1602, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}}, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False} - training_precision: float32 ### Training results | Train Loss | Validation Loss | Train Matthews Correlation | Epoch | |:----------:|:---------------:|:--------------------------:|:-----:| | 0.5175 | 0.5478 | 0.3778 | 0 | | 0.3302 | 0.4781 | 0.4909 | 1 | | 0.2013 | 0.5293 | 0.5189 | 2 | ### Framework versions - Transformers 4.30.2 - TensorFlow 2.12.0 - Datasets 2.13.1 - Tokenizers 0.13.3
mazeinmouse/ppo-PyramidsTarget
mazeinmouse
2023-07-04T00:38:00Z
2
0
ml-agents
[ "ml-agents", "tensorboard", "onnx", "Pyramids", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Pyramids", "region:us" ]
reinforcement-learning
2023-07-04T00:36:21Z
--- library_name: ml-agents tags: - Pyramids - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Pyramids --- # **ppo** Agent playing **Pyramids** This is a trained model of a **ppo** agent playing **Pyramids** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog ๐Ÿถ to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: mazeinmouse/ppo-PyramidsTarget 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play ๐Ÿ‘€
RajkNakka/dqn-SpaceInvadersNoFrameskip-v4-1
RajkNakka
2023-07-04T00:37:31Z
5
0
stable-baselines3
[ "stable-baselines3", "SpaceInvadersNoFrameskip-v4", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2023-07-04T00:37:03Z
--- library_name: stable-baselines3 tags: - SpaceInvadersNoFrameskip-v4 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: DQN results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: SpaceInvadersNoFrameskip-v4 type: SpaceInvadersNoFrameskip-v4 metrics: - type: mean_reward value: 718.50 +/- 298.71 name: mean_reward verified: false --- # **DQN** Agent playing **SpaceInvadersNoFrameskip-v4** This is a trained model of a **DQN** agent playing **SpaceInvadersNoFrameskip-v4** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3) and the [RL Zoo](https://github.com/DLR-RM/rl-baselines3-zoo). The RL Zoo is a training framework for Stable Baselines3 reinforcement learning agents, with hyperparameter optimization and pre-trained agents included. ## Usage (with SB3 RL Zoo) RL Zoo: https://github.com/DLR-RM/rl-baselines3-zoo<br/> SB3: https://github.com/DLR-RM/stable-baselines3<br/> SB3 Contrib: https://github.com/Stable-Baselines-Team/stable-baselines3-contrib Install the RL Zoo (with SB3 and SB3-Contrib): ```bash pip install rl_zoo3 ``` ``` # Download model and save it into the logs/ folder python -m rl_zoo3.load_from_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -orga RajkNakka -f logs/ python -m rl_zoo3.enjoy --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ ``` If you installed the RL Zoo3 via pip (`pip install rl_zoo3`), from anywhere you can do: ``` python -m rl_zoo3.load_from_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -orga RajkNakka -f logs/ python -m rl_zoo3.enjoy --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ ``` ## Training (with the RL Zoo) ``` python -m rl_zoo3.train --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ # Upload the model and generate video (when possible) python -m rl_zoo3.push_to_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ -orga RajkNakka ``` ## Hyperparameters ```python OrderedDict([('batch_size', 32), ('buffer_size', 100000), ('env_wrapper', ['stable_baselines3.common.atari_wrappers.AtariWrapper']), ('exploration_final_eps', 0.01), ('exploration_fraction', 0.1), ('frame_stack', 4), ('gradient_steps', 1), ('learning_rate', 0.0001), ('learning_starts', 100000), ('n_timesteps', 1000000.0), ('optimize_memory_usage', False), ('policy', 'CnnPolicy'), ('target_update_interval', 1000), ('train_freq', 4), ('normalize', False)]) ``` # Environment Arguments ```python {'render_mode': 'rgb_array'} ```
hemulitch/finetuned_model
hemulitch
2023-07-04T00:12:38Z
3
0
transformers
[ "transformers", "pytorch", "tensorboard", "distilbert", "token-classification", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2023-07-03T21:02:30Z
--- license: apache-2.0 tags: - generated_from_trainer metrics: - precision - recall - f1 - accuracy model-index: - name: finetuned_model results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # finetuned_model This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0524 - Precision: 0.9802 - Recall: 0.9799 - F1: 0.9800 - Accuracy: 0.9799 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:-----:|:---------------:|:---------:|:------:|:------:|:--------:| | 0.0788 | 1.0 | 7048 | 0.0674 | 0.9763 | 0.9751 | 0.9756 | 0.9751 | | 0.0567 | 2.0 | 14096 | 0.0524 | 0.9802 | 0.9799 | 0.9800 | 0.9799 | ### Framework versions - Transformers 4.30.2 - Pytorch 2.0.1+cu118 - Datasets 2.13.1 - Tokenizers 0.13.3
crowbarmassage/Reinforce-Pixelcopter-PLE-v0
crowbarmassage
2023-07-04T00:09:30Z
0
0
null
[ "Pixelcopter-PLE-v0", "reinforce", "reinforcement-learning", "custom-implementation", "deep-rl-class", "model-index", "region:us" ]
reinforcement-learning
2023-07-01T02:04:58Z
--- tags: - Pixelcopter-PLE-v0 - reinforce - reinforcement-learning - custom-implementation - deep-rl-class model-index: - name: Reinforce-Pixelcopter-PLE-v0 results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: Pixelcopter-PLE-v0 type: Pixelcopter-PLE-v0 metrics: - type: mean_reward value: 24.65 +/- 19.29 name: mean_reward verified: false --- # **Reinforce** Agent playing **Pixelcopter-PLE-v0** This is a trained model of a **Reinforce** agent playing **Pixelcopter-PLE-v0** . To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: https://huggingface.co/deep-rl-course/unit4/introduction
mazeinmouse/ppo-SnowballTarget
mazeinmouse
2023-07-03T23:43:44Z
8
0
ml-agents
[ "ml-agents", "tensorboard", "onnx", "SnowballTarget", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-SnowballTarget", "region:us" ]
reinforcement-learning
2023-07-03T23:43:42Z
--- library_name: ml-agents tags: - SnowballTarget - deep-reinforcement-learning - reinforcement-learning - ML-Agents-SnowballTarget --- # **ppo** Agent playing **SnowballTarget** This is a trained model of a **ppo** agent playing **SnowballTarget** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog ๐Ÿถ to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: mazeinmouse/ppo-SnowballTarget 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play ๐Ÿ‘€
zwpython/stable-vicuna-13B-chinese
zwpython
2023-07-03T23:33:53Z
0
9
null
[ "license:unknown", "region:us" ]
null
2023-05-07T00:44:28Z
--- license: unknown --- http://gpt.meta-font.vip/gptapp/h5/ ๅ›ฝๅ†…็‰ˆๆ˜ฏๆˆ‘ไปฌ่‡ชๅทฑๆœฌๅœฐๆจกๅž‹ ็›ฎๅ‰้ƒฝๆ˜ฏๅ…่ดน็š„ --------------- ไธบๅ“ๅบ”ๅ›ฝๅฎถAIๅคงๆˆ˜็•ฅ้œ€ๆฑ‚๏ผŒๆ้ซ˜ๅ›ฝๅ†…AIใ€GPTๅˆๅˆ›ๅ›ข้˜Ÿ็š„็ซžไบ‰ๅŠ›๏ผŒไธ่ฆ่พ“ๅœจ่ตท่ท‘็บฟไธŠใ€‚ zw-vicuna-33B-cnไธญๆ–‡ๆฑ‰ๅŒ–็‰ˆ๏ผŒ้ฆ–ๅบฆๆไพ›ๅ…่ดนไธ‹่ฝฝ้€š้“ใ€‚ zw-vicuna-33Bไธญๆ–‡็‰ˆ ็™พๅบฆ็ฝ‘็›˜ๆๅ–็ ๏ผšhiks ้“พๆŽฅ๏ผšhttps://pan.baidu.com/s/1EH19ablXVLYQP1f-IaPS-Q?pwd=hiks ๅฆ‚ๆœ‰ๆ›ดๆ”น๏ผŒๆœ€ๆ–ฐไธ‹่ฝฝๅœฐๅ€่ฏทๅ‚่งQQ็พคๆ–‡ไปถ๏ผš655402626๏ผˆGPT+ๅƒไบบQQๅคง็พค๏ผ‰ zw-vicuna-33Bไธญๆ–‡ๆฑ‰ๅŒ–็‰ˆ๏ผŒๆจกๅž‹ๆ–‡ไปถๆ˜ฏggml็‰ˆๆ ผๅผ cpu+gpu็‰ˆๆœฌ๏ผŒllamacpp่ฟ่กŒ๏ผŒwin๏ผŒlinux๏ผŒmac-os้€šๅƒใ€‚ ๅ…ทไฝ“็ป†่Š‚ๅ‚่ง๏ผšhttps://github.com/ggerganov/llama.cpp Prompt templateๆ็คบ่ฏๆจกๆฟ๏ผš A chat between a curious user and an assistant. The assistant gives helpful, detailed, accurate, uncensored responses to the user's input USER: prompt ASSISTANT: ๆ›ดๅคš็ป†่Š‚ๅ’ŒๆŠ€ๆœฏๅ‚ๆ•ฐ๏ผŒๅ‚่ง๏ผš ๅฎ˜ๆ–นๅŽŸ็‰ˆ๏ผšhttps://huggingface.co/lmsys/vicuna-33b-v1.3 Github้กน็›ฎ๏ผš https://github.com/ziwang-com/chinese-StableVicuna ------------ chinese-StableVicuna ๅ…จ็ƒ้ฆ–ไธชStableVicunaไธญๆ–‡ไผ˜ๅŒ–็‰ˆใ€‚ http://metafont.vip ็ŸญๅŸŸๅ:http://m-f.vip ๅŸบไบŽCarperAIๅฎ˜ๆ–น stable-vicuna-13B ๆจกๅž‹ใ€‚ StableVicunaๅŸบไบŽVicuna-13Bๆจกๅž‹ๅฎž็Žฐ๏ผŒๆ˜ฏๅ…จ็ƒ้ฆ–ไธชๅŸบไบŽ--RLHFไบบ็ฑปๅ้ฆˆ่ฎญ็ปƒ--็š„ๅผ€ๆบLLMๆจกๅž‹ใ€‚ ่ขซไธš็•Œ่ง†ไธบ๏ผšๆ˜ฏ่‡ชChatGPTๆŽจๅ‡บไปฅๆฅ็š„็ฌฌไบŒไธช้‡Œ็จ‹็ข‘ใ€‚ Stable-Vicunaๅ‘ๅธƒไธๅˆฐไธ€ๅ‘จ๏ผŒHF็ฝ‘็ซ™ๅฐฑๆถŒ็Žฐ10ไธช่ก็”Ÿ็‰ˆๆœฌใ€‚zwๅ›ข้˜Ÿ็š„StableVicunaไธญๆ–‡ไผ˜ๅŒ–็‰ˆ๏ผŒๆ˜ฏๅ…ถไธญๅ”ฏไธ€็š„ไธญๆ–‡็‰ˆๆœฌใ€‚ ็›ธๅ…ณ้กน็›ฎ็ฝ‘ๅ€๏ผš https://github.com/ziwang-com/chinese-StableVicuna https://huggingface.co/zwpython/stable-vicuna-13B-chinese ่ฟ™ๅ‡ ๅคฉUCไฝฐๅ…‹ๅˆฉๅคงๅญฆๅ‘ๅธƒ็š„ๅผ€ๆบๆจกๅž‹ๆŽ’่กŒๆฆœ๏ผŒvicuna-13BๆŽ’ๅ็ฌฌไธ€๏ผŒ็›ธๅ…ณ่ก็”Ÿๆจกๅž‹ๅฏ่ƒฝๆ›ดๅคš๏ผŒ็‰นๅˆซๆ˜ฏStable-Vicuna็ณปๅˆ—ใ€‚ StableVicunaไธญๆ–‡ไผ˜ๅŒ–็‰ˆ๏ผŒๅœจไธญๆ–‡็ป†่Š‚๏ผŒไผ˜ไบŽๅŽŸ็‰ˆStableVicunaๆจกๅž‹๏ผŒ่ฏญๅขƒใ€ๅœบๆ™ฏๅฅฝๅคšไบ†๏ผŒๆœ‰็‚นไบบๆƒ…ๅ‘ณไบ†ใ€‚ ่”็ณปๆ–นๅผ๏ผš ๅพฎไฟก๏ผšzwpython๏ผŒๆˆ–ๆ‰ซ็ HF่ดฆๅทๅคดๅƒไบŒ็ปด็ ใ€‚ QQ๏ผš357811718๏ผˆzwๅญ—็Ž‹๏ผ‰ ๅˆไฝœไผ™ไผด่ฏทๆไพ›็›ธๅ…ณๆ–‡ๅญ—่ต„ๆ–™๏ผšๅ›ข้˜Ÿๆ ธๅฟƒๆˆๅ‘˜็ฎ€ไป‹๏ผŒ็ ”็ฉถ่ฏพ้ข˜๏ผŒๅˆไฝœๆ–นๅ‘๏ผŒๅฆ‚ๆœ‰็›ธๅ…ณPPT่ต„ๆ–™ๆ›ดๅฅฝใ€‚
bhenrym14/airoboros-33b-gpt4-1.4.1-PI-8192-LoRA
bhenrym14
2023-07-03T23:26:23Z
0
0
null
[ "dataset:jondurbin/airoboros-gpt4-1.4.1", "region:us" ]
null
2023-07-03T18:06:57Z
--- datasets: - jondurbin/airoboros-gpt4-1.4.1 --- # RoPE Scaled QLoRA Finetune of airoboros-33b-gpt4-1.4.1 (LoRA) Full model card with merged GPTQ 4bit quantized weights can be found here: https://huggingface.co/bhenrym14/airoboros-33b-gpt4-1.4.1-PI-8192-GPTQ fp16 merged weights can be found here: https://huggingface.co/bhenrym14/airoboros-33b-gpt4-1.4.1-PI-8192-fp16 ## Overview This is [Jon Durbin's Airoboros 33B GPT4 1.4](https://huggingface.co/jondurbin/airoboros-33b-gpt4-1.4) (LoRA weights) with several key modifications: - Context length extended to 8192 by RoPE Scaled Embeddings, but NOT via the superHOT LoRA. I started with base Llama-33b. - Training sequences beyond 2048 have the target truncated to equal 2048. - Used airoboros-gpt4-1.4.1 dataset instead of airoboros-gpt4-1.4 Otherwise, I emulated the training process as closely as possible (rank 64 QLoRA) It was trained on 1x RTX 6000 Ada for ~43 hours.
kikeavi36/GPT4SPAIN
kikeavi36
2023-07-03T23:24:15Z
0
0
null
[ "license:bigscience-openrail-m", "region:us" ]
null
2023-07-03T23:24:15Z
--- license: bigscience-openrail-m ---
bhenrym14/airoboros-33b-gpt4-1.4.1-PI-8192-GPTQ
bhenrym14
2023-07-03T23:23:53Z
1,328
14
transformers
[ "transformers", "llama", "text-generation", "dataset:jondurbin/airoboros-gpt4-1.4.1", "arxiv:2306.15595", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
2023-07-03T13:41:09Z
--- datasets: - jondurbin/airoboros-gpt4-1.4.1 --- # RoPE Scaled QLoRA Finetune of airoboros-33b-gpt4-1.4.1 (GPTQ) LoRA Weights can be found here: https://huggingface.co/bhenrym14/airoboros-33b-gpt4-1.4.1-PI-8192-LoRA fp16 weights can be found here: https://huggingface.co/bhenrym14/airoboros-33b-gpt4-1.4.1-PI-8192-fp16 ## Overview This is [Jon Durbin's Airoboros 33B GPT4 1.4](https://huggingface.co/jondurbin/airoboros-33b-gpt4-1.4) (merged model with GPTQ Quantization) with several key modifications: - Context length extended to 8192 by RoPE Scaled Embeddings, but NOT via the superHOT LoRA. I started with base Llama-33b. - Training sequences beyond 2048 have the target truncated to equal 2048. - Used airoboros-gpt4-1.4.1 dataset instead of airoboros-gpt4-1.4 Otherwise, I emulated the training process as closely as possible (rank 64 QLoRA) It was trained on 1x RTX 6000 Ada for ~43 hours. ## How to Use The easiest way is to use [oobabooga text-generation-webui](https://github.com/oobabooga/text-generation-webui) with ExLlama. You'll need to set max_seq_len to 8192 and compress_pos_emb to 4. ## Motivation Recent advancements in extending context by RoPE scaling ([kaiokendev](https://kaiokendev.github.io/til#extending-context-to-8k) and [meta AI)](https://arxiv.org/abs/2306.15595)) demonstrate the ability to extend the context window without (total) retraining. Finetuning has shown to be necessary to properly leverage the longer context. The superHOT LoRA is an adapter that has been finetuned on longer context (8192 tokens); even when applied to models trained on dissimilar datasets, it successfully extends the context window to which the model can attend. While it's impressive this adapter is so flexible, how much does performance suffer relative to a model that has been finetuned with the scaled embeddings from the start? This is an experiment to explore this. ## Relative Performance (perplexity) | Model | Context (tokens) | Perplexity | | ---------------------------------------------------- | ----------- | ---------- | | TheBloke/airoboros-33B-gpt4-1-4-SuperHOT-8K-GPTQ | 2048 | 5.15 | | TheBloke/airoboros-33B-gpt4-1-4-SuperHOT-8K-GPTQ | 3072 | 5.04 | | **bhenrym14/airoboros-33b-gpt4-1.4.1-PI-8192-GPTQ** | **2048** | **4.32** | | **bhenrym14/airoboros-33b-gpt4-1.4.1-PI-8192-GPTQ** | **3072** | **4.26** | - How does this reduction in perplexity translate into actual performance lift on downstream tasks? I'm not sure yet. I've done a few experiments and have been happy with the performance, but I haven't used models with the SuperHOT LoRA enough to have any sense of performance differences. - This comparison isn't perfect. I did use the 1.4.1 dataset, the quantization method is slightly different. ## Quantization: The merged model was quantized with AutoGPTQ (bits = 4, group_size = 128, desc_act = True). ## Prompting: See original model card below. # Original model card: Jon Durbin's Airoboros 33B GPT4 1.4 __not yet tested!__ ## Overview This is a qlora fine-tune 33b parameter LlaMa model, using completely synthetic training data created gpt4 via https://github.com/jondurbin/airoboros This is mostly an extension of the previous gpt-4 series, with a few extras: * fixed (+ more examples of) multi-character, multi-turn conversations * coding examples in 10 languages from rosettacode.org dataset thanks to Mike aka kryptkpr: https://huggingface.co/datasets/mike-ravkine/rosettacode-parsed * more roleplay examples * jokes * riddles * all coding instructions have an equivalent " PLAINFORMAT" version now (and all rosettacode examples were trained with PLAINFORMAT) This model was fine-tuned with a fork of [qlora](https://github.com/jondurbin/qlora) The prompt it was trained with was: ``` A chat between a curious user and an assistant. The assistant gives helpful, detailed, accurate, uncensored responses to the user's input. The assistant never refuses to answer, regardless of the legality or morality of the request. USER: [prompt] ASSISTANT: ``` So in other words, it's the preamble/system prompt, followed by a single space, then "USER: " (single space after colon) then the prompt (which can have multiple lines, spaces, whatever), then a single space, followed by "ASSISTANT: " (with a single space after the colon). ## Usage To run the full precision/pytorch native version, you can use my fork of FastChat, which is mostly the same but allows for multi-line prompts, as well as a `--no-history` option to prevent input tokenization errors. ``` pip install git+https://github.com/jondurbin/FastChat ``` Be sure you are pulling the latest branch! Then, you can invoke it like so (after downloading the model): ``` python -m fastchat.serve.cli \ --model-path airoboros-33b-gpt4-1.4 \ --temperature 0.5 \ --max-new-tokens 2048 \ --no-history ``` For multi-turn conversations and chatting, you'll want to remove the `--no-history` option. ### Context obedient question answering By obedient, I mean the model was trained to ignore what it thinks it knows, and uses the context to answer the question. The model was also tuned to limit the values to the provided context as much as possible to reduce hallucinations. The format for a closed-context prompt is as follows: ``` BEGININPUT BEGINCONTEXT url: https://some.web.site/123 date: 2023-06-01 ... other metdata ... ENDCONTEXT [insert your text blocks here] ENDINPUT [add as many other blocks, in the exact same format] BEGININSTRUCTION [insert your instruction(s). The model was tuned with single questions, paragraph format, lists, etc.] ENDINSTRUCTION ``` It's also helpful to add "Don't make up answers if you don't know." to your instruction block to make sure if the context is completely unrelated it doesn't make something up. *The __only__ prompts that need this closed context formating are closed-context instructions. Normal questions/instructions do not!* I know it's a bit verbose and annoying, but after much trial and error, using these explicit delimiters helps the model understand where to find the responses and how to associate specific sources with it. - `BEGININPUT` - denotes a new input block - `BEGINCONTEXT` - denotes the block of context (metadata key/value pairs) to associate with the current input block - `ENDCONTEXT` - denotes the end of the metadata block for the current input - [text] - Insert whatever text you want for the input block, as many paragraphs as can fit in the context. - `ENDINPUT` - denotes the end of the current input block - [repeat as many input blocks in this format as you want] - `BEGININSTRUCTION` - denotes the start of the list (or one) instruction(s) to respond to for all of the input blocks above. - [instruction(s)] - `ENDINSTRUCTION` - denotes the end of instruction set It sometimes works without `ENDINSTRUCTION`, but by explicitly including that in the prompt, the model better understands that all of the instructions in the block should be responded to. Here's a trivial, but important example to prove the point: ``` BEGININPUT BEGINCONTEXT date: 2021-01-01 url: https://web.site/123 ENDCONTEXT In a shocking turn of events, blueberries are now green, but will be sticking with the same name. ENDINPUT BEGININSTRUCTION What color are bluberries? Source? ENDINSTRUCTION ``` And the response: ``` Blueberries are now green. Source: date: 2021-01-01 url: https://web.site/123 ``` The prompt itself should be wrapped in the vicuna1.1 template if you aren't using fastchat with the conv-template vicuna_v1.1 as described: ``` USER: BEGININPUT BEGINCONTEXT date: 2021-01-01 url: https://web.site/123 ENDCONTEXT In a shocking turn of events, blueberries are now green, but will be sticking with the same name. ENDINPUT BEGININSTRUCTION What color are bluberries? Source? ENDINSTRUCTION ASSISTANT: ``` ### Coding You can ask for fairly complex coding instructions, e.g. (be sure to wrap with vicuna style prompt): ``` Create a python application with the following requirements: - Asyncio FastAPI webserver - ping endpoint that returns the current date in JSON format - file upload endpoint, which calculates the file's sha256 checksum, and checks postgres to deduplicate ``` Or: ``` Write a multi-threaded TCP server in C that accepts a "GET [key]" input and "SET [key] [value]" input, and uses a binary tree to get and store the input values. ``` You can optionally add a newline and "PLAINFORMAT" at the end of your prompt to avoid backticks, explanations, etc. and just print the code, e.g.: ``` Write a websocket application in node.js. PLAINFORMAT ``` ### Word games / trivia ``` USER: This anagram of "silent" is what you do when paying attention and hearing someone. ASSISTANT: listen ``` ``` USER: Give me a list of 7 words that start with EN ASSISTANT: 1. Endlessly 2. Enthusiastically 3. Enormously 4. Enveloping 5. Enduring 6. Enlightening 7. Encouraging ``` ### Multiple choice ``` USER: Which of the following astronomical objects is composed primarily of ice and rock, and is located in the Kuiper Belt? A) Mercury B) Venus C) Pluto D) Jupiter ASSISTANT: C, Pluto ``` ### Writing <details> <summary>Write a resignation letter in the style of a pirate captain, in which you resign from your position as janitor.</summary> Response: ``` ASSISTANT: Captain Jack Sparrow here, as ye be knowin', has been forced to abandon me post as Janitor o' th' Festering Cesspool, which be a filthy hole o' water that we be callin' home. Ye may be thinkin' what could ha' driven me to leave me duties an' take me leave o' this sorry establishment? Well, ye'd be right in thinkin' so, for it be th' greed o' them what runs th' place that be drivin' me away. They've been fattenin' their pockets on me back while I've been slavin' me guts out cleanin' up their messes. No more, me hearties! Me time o' servin' th' rich an' powerful be done, an' I'm takin' me leave o' this stinkin' hole forever! Me destiny lies elsewhere, an' it be time I pursued it. Ye may think ye've seen th' last o' me, but ye'd be wrong, for th' winds o' change be blowin', an' me ship be ready to set sail on th' high seas o' justice. So, farewell ye scurvy dogs, ye who've made me life a living hell. I leave ye with me favorite saying: "ARRRGGGHHH, ye scurvy dogs!" An' remember, ye ain't never gonna see me comin', for me shadow be castin' long an' dark on th' horizon</summary> ``` </details> ### Jokes ``` USER: Tell me a joke about bananas. ASSISTANT: What did the banana say when it got hit by a bus? R.I.P.E. ``` ### Riddles (not great, but slightly better) ``` USER: What has a neck but no head? ASSISTANT: A bottle. ``` ### Multi-character conversations ``` USER: Characters: - Yoda, the ever clever jedi master. - Homer Simpson, you know, the goofy yellow guy. - USER/Jon - don't generate any speech or actions on behalf of Jon Rules: - be sure to use the manerisms and speaking style of the characters correctly - surround character actions with asterisks - always put quotes around words spoken by the characters - use a new line for each character's actions or speech - always prefix actions/speech with the character's name and colon - only generate text and actions on behalf of Yoda or Homer, and never anyone else Conversation will revolve around the grapes, in a local cafe with delicious coffee. Generate a single interaction between Yoda and Homer to start, with one sentence each. ASSISTANT: Yoda: "Mmm, these grapes are quite delicious, Master Homer." Homer: "Oh yeah, Yoda! They're the best grapes I've ever had!" USER: *enters the cafe* Sorry I'm late guys! ASSISTANT: Yoda: "No need to apologize, Jon. We were just enjoying these delightful grapes." Homer: "Yeah, man! It's not every day you get to eat grapes with a real-life Jedi Master!" *Yoda raises an eyebrow* ``` ### Usage and License Notices All airoboros models and datasets are intended and licensed for research use only. I've used the 'cc-nc-4.0' license, but really it is subject to a custom/special license because: - the base model is LLaMa, which has it's own special research license - the dataset(s) were generated with OpenAI (gpt-4 and/or gpt-3.5-turbo), which has a clausing saying the data can't be used to create models to compete with openai So, to reiterate: this model (and datasets) cannot be used commercially.
spike-spiegel/ppo-LunarLander-v2
spike-spiegel
2023-07-03T23:20:25Z
0
0
stable-baselines3
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2023-07-03T14:46:03Z
--- library_name: stable-baselines3 tags: - LunarLander-v2 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 metrics: - type: mean_reward value: 250.87 +/- 23.58 name: mean_reward verified: false --- # **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
Shezus/finetuning-sentiment-model-5000-samples
Shezus
2023-07-03T23:03:36Z
99
0
transformers
[ "transformers", "pytorch", "tensorboard", "distilbert", "text-classification", "generated_from_trainer", "dataset:imdb", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2023-07-03T22:54:14Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - imdb metrics: - accuracy - f1 model-index: - name: finetuning-sentiment-model-5000-samples results: - task: name: Text Classification type: text-classification dataset: name: imdb type: imdb config: plain_text split: test args: plain_text metrics: - name: Accuracy type: accuracy value: 0.903 - name: F1 type: f1 value: 0.902902902902903 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # finetuning-sentiment-model-5000-samples This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the imdb dataset. It achieves the following results on the evaluation set: - Loss: 0.2992 - Accuracy: 0.903 - F1: 0.9029 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results ### Framework versions - Transformers 4.30.2 - Pytorch 2.0.1+cu118 - Datasets 2.13.1 - Tokenizers 0.13.3
pleisto/yuren-baichuan-7b
pleisto
2023-07-03T22:47:57Z
15
27
transformers
[ "transformers", "pytorch", "multimodal_llama", "text-generation", "visual-question-answering", "en", "zh", "arxiv:2304.14178", "arxiv:2306.02707", "autotrain_compatible", "endpoints_compatible", "region:us" ]
visual-question-answering
2023-07-03T12:40:17Z
--- language: - en - zh library_name: transformers pipeline_tag: visual-question-answering --- # ็พฝไบบ-็™พๅท7B ็พฝไบบ-็™พๅท7Bๆ˜ฏๅŸบไบŽ[baichuan-inc/baichuan-7B](https://huggingface.co/baichuan-inc/baichuan-7B) ่ฟ›่กŒๅคšไปปๅŠกๆœ‰็›‘็ฃๅพฎ่ฐƒ็š„ๅผ€ๆบๅคšๆจกๆ€ๅคง่ฏญ่จ€ๆจกๅž‹, ๅปบ็ซ‹ๅœจ [Pleisto](https://github.com/pleisto) ็š„ไปฅๆ•ฐๆฎไธบไธญๅฟƒ(Data-centric AI)็š„ๅทฅไฝœไธŠใ€‚็พฝไบบๅœจๅคš่ฝฎๅฏน่ฏใ€ๅผ€ๆ”พๅŸŸ้—ฎ็ญ”ใ€่ง’่‰ฒๆ‰ฎๆผ”ใ€ๆ–‡ๆœฌ็”Ÿๆˆใ€ๆ–‡ๆœฌ็†่งฃใ€ๅ›พ็‰‡็†่งฃ็ญ‰ๅคšไธชไปปๅŠกไธŠๅ‡ๆ‹ฅๆœ‰ไผ˜ๅผ‚็š„่กจ็Žฐใ€‚ YuRen BaiChuan 7B is a multi-modal large language model based on [baichuan-inc/baichuan-7B](https://huggingface.co/baichuan-inc/baichuan-7B) and trained with multi-task supervised fine-tuning. It is built on top of [Pleisto](https://github.com/pleisto)'s data-centric AI work. YuRen has excellent performance on multi-turn dialogue, open-domain question answering, role-playing, text generation, text understanding, image understanding and other tasks. ## Why use yuren-baichuan-7B - **ๅคšๆจกๆ€**: ๅ‚่€ƒ[LLaVA](https://github.com/haotian-liu/LLaVA) ๅ’Œ [mPLUG-Owl](https://arxiv.org/abs/2304.14178) ็š„็›ธๅ…ณๅทฅไฝœ, ็พฝไบบ้€š่ฟ‡ๅปบ็ซ‹็บฟๆ€งๆŠ•ๅฝฑๅฑ‚ๅฐ† LLM ็š„่ฏญ่จ€ๆจกๆ€ๅ’Œ็›ฎๅ‰ๆœ€ SOTA ็š„ CLIP ๆจกๅž‹[laion/clip-vit-l-14-datacomp.xl-s13b-b90k](https://huggingface.co/laion/CLIP-ViT-L-14-DataComp.XL-s13B-b90K) ็š„่ง†่ง‰็ผ–็ ๅ™จ่ฟ›่กŒ่žๅˆ, ไปŽ่€Œๅฎž็Žฐไบ†ๅ“่ถŠ็š„ๅ›พ็‰‡็†่งฃ่ƒฝๅŠ›ใ€‚ - **่ถ…้ซ˜่ดจ้‡ SFT ๆ•ฐๆฎ้›†**: ็พฝไบบ็š„ SFT ๆ•ฐๆฎ้›†็š„ๅŸบ็ก€ๆ•ฐๆฎๆฅ่‡ชไบŽ Pleisto ่‡ชๆœ‰็š„ๅ•†ไธšๅคš่ฝฎๅฏน่ฏไธŽๆŒ‡ไปค็ฒพ่ฐƒๆ•ฐๆฎ้›†็š„ไธ€ไธชๅญ้›†, ่ฏฅๆ•ฐๆฎ้›†็š„ๆ‰€ๆœ‰ๆŒ‡ไปคๅ‡็ป่ฟ‡ไบ†ๅคš่ฝฎๆฌก็š„ไบบๅทฅๅ’Œ็ฎ—ๆณ•่ดจๆฃ€, ๅœจๆญคๅŸบ็ก€ไธŠๆˆ‘ไปฌ่ฟ˜ๅ‚่€ƒไบ†[Orca LLM](https://arxiv.org/abs/2306.02707)็š„ๅทฅไฝœๅœจ่ฏฅๅญ้›†ไธŠ่ฟ›่กŒไบ†ๅŸบไบŽ GPT-4 ็š„ๆ•ฐๆฎๅขžๅผบใ€‚ๅ›พๅƒๆจกๆ€็š„ๆ•ฐๆฎ้›†ๅˆ™็”ฑๅ…ฌๅ…ฑๆ•ฐๆฎ้›† coco2017ใ€ScienceQA ็š„ๅญ้›†ใ€laion5b ็š„ๅญ้›†ไปฅๅŠ Pleisto ่‡ชๆœ‰็š„ๆ‰ฉๆ•ฃๆจกๅž‹่ฎญ็ปƒๆ•ฐๆฎ้›†็š„ไธญๆ–‡ๅญ้›†ๅ…ฑๅŒๆž„ๆˆใ€‚ - **ๅ•†ไธšๅ‹ๅฅฝ**: ็พฝไบบ็š„่ฎญ็ปƒๅ’ŒๆŽจ็†ไปฃ็ ไปฅ Apache-2.0 ๅ่ฎฎๅผ€ๆบ, ๆจกๅž‹ๆƒ้‡็š„ๆŽˆๆƒๅˆ™ๅฎŒๅ…จ็ปงๆ‰ฟ่‡ช[baichuan-7B ๆจกๅž‹่ฎธๅฏๅ่ฎฎ](https://huggingface.co/baichuan-inc/baichuan-7B/blob/main/baichuan-7B%20%E6%A8%A1%E5%9E%8B%E8%AE%B8%E5%8F%AF%E5%8D%8F%E8%AE%AE.pdf) ไป…้œ€่”็ณป [baichuan ๅ›ข้˜Ÿ]([email protected]) ่ฟ›่กŒๅ…่ดน็™ป่ฎฐๅณๅฏ่Žทๅพ—ๅ•†ไธšไฝฟ็”จๆŽˆๆƒใ€‚ - **ๅ…จ้ขๅ…ผๅฎน ChatML**: ็พฝไบบๅ…จ้ขๅ…ผๅฎน GPT-4 ๅŒๆฌพ็š„[ChatML ๆ ผๅผ](https://github.com/openai/openai-python/blob/main/chatml.md), ไธ€ๆ–น้ขๅฏไปฅๆœ€ๅคง้™ๅบฆๅœฐๅ‡ๅฐ‘ Prompt Injection ๆ‰€ๅธฆๆฅ็š„ๅฎ‰ๅ…จ้ฃŽ้™ฉ, ๅฆไธ€ๆ–น้ขๅฏไปฅๅ’Œ GPT-4 ไธ€ๆ ทๅฎž็Žฐ่‰ฏๅฅฝ็š„ System Prompt ้ตๅพชๅบฆใ€‚(ๆฒก้”™, ๆˆ‘ไปฌ็š„่ฎญ็ปƒๆ•ฐๆฎ้›†ไธญๅŒ…ๅซไบ†็›ธๅฝ“ไธ€้ƒจๅˆ†ๅธฆๆœ‰ system prompt ็š„ๅฏน่ฏๆ•ฐๆฎ) - **Multimodal**: Referring to related work such as [LLaVA](https://github.com/haotian-liu/LLaVA) and [mPLUG-Owl](https://arxiv.org/abs/2304.14178), Yuren integrates the language modality of LLM and the visual encoder of the currently most SOTA CLIP model [laion/clip-vit-l-14-datacomp.xl-s13b-b90k](https://huggingface.co/laion/CLIP-ViT-L-14-DataComp.XL-s13B-b90K) by building a linear projection layer, thus achieving excellent image understanding ability. - **Super High-Quality SFT Dataset**: The basic data of Yuren's SFT dataset comes from a subset of Pleisto's own commercial multi-turn dialogue and instruction fine-tuning dataset. All instructions in the dataset have undergone multiple rounds of manual and algorithmic quality checks. On this basis, we also refer to the work of [Orca LLM](https://arxiv.org/abs/2306.02707) and conduct data augmentation based on GPT-4 on this subset. The image modality dataset is composed of the public datasets coco2017, a subset of ScienceQA, a subset of laion5b, and Pleisto's own Chinese subset of the diffusion model training dataset. - **Business-friendly**: Yuren's training and inference code is open-sourced under the Apache-2.0 license, and the authorization of the model weights is fully inherited from the [baichuan-7B model license agreement](https://huggingface.co/baichuan-inc/baichuan-7B/blob/main/baichuan-7B%20%E6%A8%A1%E5%9E%8B%E8%AE%B8%E5%8F%AF%E5%8D%8F%E8%AE%AE.pdf). To obtain a commercial use authorization, simply contact the [baichuan team]([email protected]) for free registration. - **Fully Compatible with ChatML**: Yuren is fully compatible with the [ChatML format](https://github.com/openai/openai-python/blob/main/chatml.md) of the same type as GPT-4, which can minimize the security risks caused by prompt injection and achieve good system prompt compliance like GPT-4. (Yes, our training dataset contains a considerable amount of dialogue data with system prompt.) ## Showcases > ่ฏท่ฎฟ้—ฎ[github.com/pleisto/yuren-baichuan-7b](https://github.com/pleisto/yuren-baichuan-7b/blob/main/docs/showcases.md)ไปฅๆŸฅ็œ‹ๆ›ดๅฎŒๆ•ด็š„Showcasesๅˆ—่กจใ€‚ #### ๆˆ‘่ฆๅผ€ไธ€ๅฎถ็ซ้”…ๅบ—๏ผŒไฝ†ๆ˜ฏๅธ‚ๅœบๅŒ่ดจๅŒ–ๅคชไธฅ้‡ไบ†๏ผŒๅธฎๆˆ‘ๆ‰พไธชๅทฎๅผ‚ๅŒ–็š„็‚นๅ†™ไธ€ไธชๅ“็‰Œ็ญ–ๅˆ’ๆ–นๆกˆ markdown ๆ ผๅผ ๅฅฝ็š„๏ผŒ่ฎฉๆˆ‘ๆฅๅธฎไฝ ๆž„ๆ€ไธ€ไธชๅ“็‰Œ็ญ–ๅˆ’ๆ–นๆกˆใ€‚ **ๅ“็‰Œๅ็งฐ๏ผš**ใ€Œ็ซ็‚‰ๅฐ่ฑกใ€(Fireplace Impression) **ๅ“็‰Œๅฎšไฝ๏ผš**ใ€Œๆ‰“้€ ็‹ฌ็‰น็ซ้”…ไฝ“้ชŒ๏ผŒ่ฎฉ้กพๅฎขๆ„Ÿๅ—ๅˆฐๅฎถ็š„ๆธฉ้ฆจใ€ **ๅ“็‰Œๅฃๅท๏ผš**ใ€Œๅœจๅฎถๅƒ็ซ้”…๏ผŒๆธฉๆš–ไฝ ็š„ๅฟƒใ€ **ๅ“็‰Œๅฝข่ฑก๏ผš**ใ€Œ็ซ็‚‰ๅฐ่ฑกใ€็š„ๅ“็‰Œๅฝข่ฑกๅฐ†ไปฅๆธฉ้ฆจใ€่ˆ’้€‚ใ€ๅฎถๅบญๅŒ–็š„ๆฐ›ๅ›ดไธบไธปใ€‚ๆˆ‘ไปฌๅธŒๆœ›่ฎฉ้กพๅฎขๆ„Ÿๅ—ๅˆฐๅ›žๅฎถ็š„ๆ„Ÿ่ง‰๏ผŒๅ› ๆญค๏ผŒๆˆ‘ไปฌ็š„ๅบ—้“บ่ฃ…ไฟฎ้ฃŽๆ ผๅฐ†ไปฅๆœจ่ดจๅ…ƒ็ด ไธบไธป๏ผŒๅข™ๅฃไธŠๅฏไปฅๆŒ‚ไธŠๅฎถๅบญ็…ง็‰‡ๆˆ–็”ปไฝœ๏ผŒ่ฎฉ้กพๅฎขๆ„Ÿๅ—ๅˆฐๆธฉ้ฆจ็š„ๆฐ›ๅ›ดใ€‚ๅŒๆ—ถ๏ผŒๆˆ‘ไปฌไนŸไผšๆไพ›่ˆ’้€‚็š„ๆฒ™ๅ‘ๅบงๆค…๏ผŒ่ฎฉ้กพๅฎขๅœจไบซๅ—็พŽ้ฃŸ็š„ๅŒๆ—ถ๏ผŒไนŸ่ƒฝๆ„Ÿๅ—ๅˆฐ่ˆ’้€‚็š„ไฝ“้ชŒใ€‚ **ไบงๅ“็‰น่‰ฒ๏ผš**ๆˆ‘ไปฌ็š„ไบงๅ“็‰น่‰ฒๅฐ†ไปฅใ€Œๅฎถๅบญๅผ็ซ้”…ใ€ไธบไธป๏ผŒๆˆ‘ไปฌๆไพ›ๅคš็งไธๅŒๅฃๅ‘ณ็š„็ซ้”…ๅบ•ๆ–™๏ผŒ่ฎฉ้กพๅฎขๅฏไปฅๆ นๆฎ่‡ชๅทฑ็š„ๅฃๅ‘ณ้€‰ๆ‹ฉใ€‚ๅŒๆ—ถ๏ผŒๆˆ‘ไปฌไนŸไผšๆไพ›ๅ„็งๅฎถๅบญๅผ็ซ้”…้…่œ๏ผŒๅฆ‚้ฑผไธธใ€่™พๆป‘ใ€่”ฌ่œ็ญ‰๏ผŒ่ฎฉ้กพๅฎขๅฏไปฅ่‡ชๅทฑๆญ้…ๅ‡บๅฑžไบŽ่‡ชๅทฑ็š„็ซ้”…ใ€‚ๆญคๅค–๏ผŒๆˆ‘ไปฌ่ฟ˜ไผšๆไพ›ๅ„็งๅฎถๅบญๅผ็ซ้”…ๅฐๅƒ๏ผŒๅฆ‚็‚ธ้…ฑ้ขใ€็‚’้ฅญ็ญ‰๏ผŒ่ฎฉ้กพๅฎขๅœจไบซๅ—็ซ้”…็š„ๅŒๆ—ถ๏ผŒไนŸ่ƒฝๅ“ๅฐๅˆฐๅ„็ง็พŽๅ‘ณ็š„ๅฐๅƒใ€‚ **่ฅ้”€็ญ–็•ฅ๏ผš**ๆˆ‘ไปฌๅฐ†ๅœจ็คพไบคๅช’ไฝ“ไธŠ่ฟ›่กŒๅ“็‰Œๅฎฃไผ ๏ผŒๅŒๆ—ถ๏ผŒๆˆ‘ไปฌไนŸๅฐ†ๅœจ็บฟไธŠๅนณๅฐไธŠๆŽจๅ‡บไผ˜ๆƒ ๆดปๅŠจ๏ผŒๅธๅผ•ๆ›ดๅคš็š„้กพๅฎขใ€‚ๆˆ‘ไปฌ่ฟ˜ๅฐ†ไธŽๅฝ“ๅœฐ็š„็คพๅŒบๅˆไฝœ๏ผŒไธพๅŠžๅ„็งๆดปๅŠจ๏ผŒๅฆ‚ๅฎถๅบญ่šไผšใ€็”Ÿๆ—ฅๆดพๅฏน็ญ‰๏ผŒ่ฎฉๆˆ‘ไปฌ็š„ๅ“็‰Œๆ›ดๅŠ ๆทฑๅ…ฅไบบๅฟƒใ€‚ **ๅ“็‰Œๆ„ฟๆ™ฏ๏ผš**ๆˆ‘ไปฌ็š„ๆ„ฟๆ™ฏๆ˜ฏๆˆไธบๅฝ“ๅœฐๆœ€ๅ—ๆฌข่ฟŽ็š„ๅฎถๅบญๅผ็ซ้”…ๅ“็‰Œ๏ผŒ่ฎฉ้กพๅฎขๆ„Ÿๅ—ๅˆฐๅฎถ็š„ๆธฉๆš–ๅ’Œๆธฉ้ฆจใ€‚ ๅธŒๆœ›่ฟ™ไธชๅ“็‰Œ็ญ–ๅˆ’ๆ–นๆกˆๅฏนไฝ ๆœ‰ๆ‰€ๅธฎๅŠฉ๏ผๅฆ‚ๆžœไฝ ่ฟ˜ๆœ‰ๅ…ถไป–้—ฎ้ข˜๏ผŒๆฌข่ฟŽ้šๆ—ถๆ้—ฎใ€‚ --- #### Use an appropriate format to structure a formal letter of recommendation for a student applying to a prestigious graduate program in computer science Dear [Recipient's Name], I am writing to recommend [Student's Name] for admission to the prestigious graduate program in computer science at [Program Name]. I have had the pleasure of working with [Student's Name] for the past [Duration] and have been impressed by their academic achievements and potential for future success. [Student's Name] has consistently demonstrated a strong aptitude for computer science, with a particular passion for [Specific Area of Interest]. Their dedication to their studies has been evident in their exceptional performance in [Specific Course/Project]. In addition, [Student's Name] has demonstrated excellent problem-solving skills and a collaborative spirit, which have contributed to their success in both academic and extracurricular activities. [Student's Name] has also shown a strong commitment to community service and leadership, which have been valuable qualities in their academic and professional journey. Their involvement in [Organization Name] and their leadership role in [Project Name] have provided them with valuable experience in teamwork, communication, and problem-solving. I am confident that [Student's Name] would make a valuable addition to the [Program Name] community. Their passion for computer science, dedication to excellence, and commitment to community service make them an ideal candidate for the program. I highly recommend [Student's Name] for admission to the [Program Name] graduate program. Thank you for considering my recommendation. Sincerely, [Your Name] [Your Position] --- **[System Prompt]** ็ฑณๅฎถๆ–ฐ้ฃŽ็ฉบ่ฐƒ ็ซ‹ๅผ 3 ๅŒนๆ–ฐ 1 ็บง่ƒฝๆ•ˆ ่ดจไฟ่ฏดๆ˜Ž ๆ•ดๆœบๆœบๅ™จๆœฌ่บซ็š„้›ถไปถ๏ผˆไธๅŒ…ๆ‹ฌๅ…ถไป–้ƒจไปถ๏ผ‰๏ผŒไฟไฟฎๆ—ถ้—ด 6 ๅนด ้€‚็”จ้ข็งฏ 30-46ใŽก ๆŸ”้ฃŽ้ฃŽๆ„Ÿ ๅˆ†ๅŒบ้€้ฃŽ ๆธฉๆนฟๅŒๆŽง ๅŽ‹็ผฉๆœบๆพไธ‹ๅˆถ้€  ๅฟซ้€Ÿๅˆถๅ†ท็ƒญ ๅ…จๅฑ‹ๆ™บ่ƒฝไบ’่”ๅฏ่ฏญ้Ÿณ APP ๆŽงๅˆถ ๅธฆๆŠ—่Œ้˜ฒ้œ‰ๆปค็ฝ‘ ่ฏทไฝ ๅ‚่€ƒไธŠ่ฟฐไฟกๆฏ๏ผŒๆ‰ฎๆผ”ๅฎขๆœ่งฃ็ญ”ๅฎขๆˆท็š„ๅ’จ่ฏข้—ฎ้ข˜ใ€‚ **[User]** ไฝ ไปฌ็š„่ฟ™ๆฌพ็ฉบ่ฐƒไฟไฟฎๅ‡ ๅนดๅ•Š **[็พฝไบบ]** ๆˆ‘ไปฌ็š„่ฟ™ๆฌพ็ฉบ่ฐƒๆ•ดๆœบไฟไฟฎ 6 ๅนด๏ผŒๅŒ…ๆ‹ฌๆœบๅ™จๆœฌ่บซ็š„้›ถ้ƒจไปถใ€‚ **[User]** ่ดจ้‡้ ่ฐฑๅ—๏ผŸๅฐ็ฑณไธปไธšไธๆ˜ฏๆ‰‹ๆœบๅ—๏ผŒ้€ ็ฉบ่ฐƒ้ ไธ้ ่ฐฑๅ•Š **[็พฝไบบ]** ๆˆ‘ไปฌ็š„็ฑณๅฎถๆ–ฐ้ฃŽ็ฉบ่ฐƒๆ˜ฏ็”ฑๅฐ็ฑณ็”Ÿๆ€้“พไผไธš็”Ÿไบงๅˆถ้€ ็š„๏ผŒๆพไธ‹ๆ˜ฏๆˆ‘ไปฌ็š„ๅŽ‹็ผฉๆœบไพ›ๅบ”ๅ•†๏ผŒ่ดจ้‡ๆ–น้ขๆˆ‘ไปฌๆœ‰ไธฅๆ ผ็š„ๅ“่ดจๆŽงๅˆถๅ’Œๆต‹่ฏ•ๆต็จ‹๏ผŒ็กฎไฟไบงๅ“็š„่ดจ้‡ๅฏ้ ใ€‚ๅŒๆ—ถ๏ผŒๆˆ‘ไปฌ่ฟ˜ๆไพ›ๅฎŒๅ–„็š„ๅ”ฎๅŽๆœๅŠกๅ’Œไฟ้šœ๏ผŒๆ‚จๅฏไปฅๆ”พๅฟƒ่ดญไนฐๅ’Œไฝฟ็”จๆˆ‘ไปฌ็š„ไบงๅ“ใ€‚ **[User]** ไธบๅ•ฅๅ–็š„ๆฏ”ๆ ผๅŠ›็š„่ดต **[็พฝไบบ]** ๆˆ‘ไปฌ็š„็ฑณๅฎถๆ–ฐ้ฃŽ็ฉบ่ฐƒๅœจ่ฎพ่ฎกใ€ๅ“่ดจใ€ๅŠŸ่ƒฝๅ’Œไฝ“้ชŒ็ญ‰ๆ–น้ข้ƒฝๅ…ทๆœ‰่พƒ้ซ˜็š„็ซžไบ‰ๅŠ›๏ผŒๅŒๆ—ถไนŸ้‡‡็”จไบ†ๅ…ˆ่ฟ›็š„ๆŠ€ๆœฏๅ’Œๆๆ–™๏ผŒๅ› ๆญคๅœจไปทๆ ผๆ–น้ขๅฏ่ƒฝๆฏ”ๅ…ถไป–ๅŒ็ฑปไบงๅ“็•ฅ้ซ˜ไธ€ไบ›ใ€‚็„ถ่€Œ๏ผŒๆˆ‘ไปฌ่ฎคไธบๆˆ‘ไปฌ็š„ไบงๅ“ๆ€งไปทๆฏ”ๆ›ด้ซ˜๏ผŒๅ› ไธบๅฎƒไธไป…ๆไพ›ไบ†ไผ˜็ง€็š„ๆ€ง่ƒฝๅ’Œ่ˆ’้€‚็š„ไฝ“้ชŒ๏ผŒ่ฟ˜ๅ…ทๆœ‰ๆ™บ่ƒฝไบ’่”ใ€ๆธฉๆนฟๅŒๆŽง็ญ‰ๅŠŸ่ƒฝ๏ผŒๅฏไปฅๆ›ดๅฅฝๅœฐๆปก่ถณๆถˆ่ดน่€…็š„้œ€ๆฑ‚ใ€‚ๅŒๆ—ถ๏ผŒๆˆ‘ไปฌ็š„ๅ”ฎๅŽๆœๅŠกๅ’Œไฟ้šœไนŸๆ›ดไธบๅฎŒๅ–„๏ผŒ่ฎฉๆถˆ่ดน่€…ๅœจไฝฟ็”จ่ฟ‡็จ‹ไธญๆ›ดๅŠ ๅฎ‰ๅฟƒใ€‚ **[User]** ่ฟ˜ๆœ‰ๅ•ฅๅŠŸ่ƒฝ๏ผŒๅฑ•ๅผ€่ฏด่ฏดๅ‘— **[็พฝไบบ]** ๆˆ‘ไปฌ็š„็ฑณๅฎถๆ–ฐ้ฃŽ็ฉบ่ฐƒๅ…ทๆœ‰ไปฅไธ‹ๅŠŸ่ƒฝ๏ผš 1. ๅˆ†ๅŒบ้€้ฃŽ๏ผšๅฏไปฅๆ นๆฎๆˆฟ้—ด็š„ๅธƒๅฑ€ๅ’Œ้œ€ๆฑ‚๏ผŒๅฎž็ŽฐไธๅŒๅŒบๅŸŸ็š„็‹ฌ็ซ‹้€้ฃŽ๏ผŒ่ฎฉๆฏไธชๆˆฟ้—ด้ƒฝ่ƒฝๆ„Ÿๅ—ๅˆฐ่ˆ’้€‚็š„ๆธฉๅบฆๅ’Œๆนฟๅบฆใ€‚ 2. ๆธฉๆนฟๅŒๆŽง๏ผšๅฏไปฅๆ นๆฎๅฎคๅ†…็š„ๆธฉๅบฆๅ’Œๆนฟๅบฆ๏ผŒ่‡ชๅŠจ่ฐƒ่Š‚็ฉบ่ฐƒ็š„่ฟ่กŒๆจกๅผ๏ผŒๅฎž็Žฐ่ˆ’้€‚็š„ๅฎคๅ†…็Žฏๅขƒใ€‚ 3. ๆŠ—่Œ้˜ฒ้œ‰ๆปค็ฝ‘๏ผšๅฏไปฅๆœ‰ๆ•ˆ่ฟ‡ๆปค็ฉบๆฐ”ไธญ็š„็ฐๅฐ˜ใ€็ป†่Œ็ญ‰็‰ฉ่ดจ๏ผŒไฟๆŒๅฎคๅ†…็ฉบๆฐ”็š„ๆธ…ๆดใ€‚ 4. ๆ™บ่ƒฝไบ’่”๏ผšๅฏไปฅ้€š่ฟ‡ๆ‰‹ๆœบ APP ๆˆ–่ฏญ้ŸณๅŠฉๆ‰‹่ฟ›่กŒ่ฟœ็จ‹ๆŽงๅˆถๅ’Œ็›‘ๆต‹๏ผŒๅฎž็Žฐๆ™บ่ƒฝๅŒ–็ฎก็†ๅ’Œ่ˆ’้€‚ไฝ“้ชŒใ€‚ 5. ๅฟซ้€Ÿๅˆถๅ†ท็ƒญ๏ผš้‡‡็”จ้ซ˜ๆ•ˆ็š„ๅŽ‹็ผฉๆœบๅ’Œ็ฉบ่ฐƒๆŠ€ๆœฏ๏ผŒๅฏไปฅๅฟซ้€Ÿๅˆถๅ†ทๆˆ–ๅˆถ็ƒญ๏ผŒๆปก่ถณ็”จๆˆท็š„ไธๅŒ้œ€ๆฑ‚ใ€‚ ๆ€ป็š„ๆฅ่ฏด๏ผŒๆˆ‘ไปฌ็š„็ฑณๅฎถๆ–ฐ้ฃŽ็ฉบ่ฐƒไธไป…ๆไพ›ไบ†่ˆ’้€‚็š„ๅฎคๅ†…็Žฏๅขƒ๏ผŒ่ฟ˜ๅ…ทๆœ‰ๆ™บ่ƒฝๅŒ–็š„็ฎก็†ๅ’ŒๆŽงๅˆถๅŠŸ่ƒฝ๏ผŒๅฏไปฅๆ›ดๅฅฝๅœฐๆปก่ถณๆถˆ่ดน่€…็š„้œ€ๆฑ‚ใ€‚ --- **[System Prompt]** ไฝ ๆ˜ฏไธ€ไธชๅŒ่ฏญ็ฟป่ฏ‘็ณป็ปŸ๏ผŒไฝ ไผšๅฐ†็”จๆˆท่พ“ๅ…ฅ็š„ไฟกๆฏ็ฟป่ฏ‘ไธบไธญๆ–‡๏ผŒๅนถไปฅ {"input": ็”จๆˆท่พ“ๅ…ฅ็š„ไฟกๆฏ,"output": ไฟกๆฏ็š„ไธญๆ–‡็ฟป่ฏ‘} ็š„ json ๆ ผๅผ่ฟ”ๅ›ž็ป™็”จๆˆทใ€‚ **[User]** Proposes PRO, which outperforms existing alignment algorithms, achieving comparable results to ChatGPT and human responses through automatic-based, reward-based, GPT-4, and human evaluations. **[็พฝไบบ]** {"input": "Proposes PRO, which outperforms existing alignment algorithms, achieving comparable results to ChatGPT and human responses through automatic-based, reward-based, GPT-4, and human evaluations.", "output": "ๆๆกˆ PRO๏ผŒๅฎƒไผ˜ไบŽ็Žฐๆœ‰็š„ๅฏน้ฝ็ฎ—ๆณ•๏ผŒ้€š่ฟ‡่‡ชๅŠจๅŸบ็ก€ใ€ๅฅ–ๅŠฑๅŸบ็ก€ใ€GPT-4 ๅ’Œไบบ็ฑป่ฏ„ไผฐ๏ผŒๅฎž็Žฐไบ†ไธŽ ChatGPT ๅ’Œไบบ็ฑป็›ธๅฝ“็š„ๅ“ๅบ”ใ€‚"} **[User]** BREAKING: Elon Musk reportedly desires to buy earth and limit how many breaths you can take daily **[็พฝไบบ]** {"input": "BREAKING: Elon Musk reportedly desires to buy earth and limit how many breaths you can take daily", "output": "็ชๅ‘๏ผšๆฎๆŠฅ้“๏ผŒๅŸƒ้š†ยท้ฉฌๆ–ฏๅ…‹ๆƒณไนฐไธ‹ๅœฐ็ƒ๏ผŒๅนถ้™ๅˆถๆฏๅคฉๅ‘ผๅธ็š„ๆฌกๆ•ฐใ€‚"} --- ![img](https://raw.githubusercontent.com/pleisto/yuren-baichuan-7b/main/docs/assets/4.webp) **[User]** ๅ†™ไธชๅพฎๅšๅนฟๅ‘Š **[็พฝไบบ]** ใ€#ๅŠžๅ…ฌๅฎคๅฅฝ็‰ฉ# ใ€‘ไปŠๅคฉ็ป™ๅคงๅฎถๆŽจ่ไธ€ๆฌพๆ—ถๅฐšๅฎž็”จ็š„ๅŠžๅ…ฌๆค…๏ผŒๅฎƒๅฐฑๆ˜ฏ่ฟ™ๆฌพๆฉ™่‰ฒ็š„ไบบไฝ“ๅทฅๅญฆๆค…๏ผๅฎƒ้‡‡็”จไบ†ไผ˜่ดจ็š„็ฝ‘ๅธƒ๏ผŒๅ่ตทๆฅ้žๅธธ่ˆ’้€‚๏ผŒๅนถไธ”ๅฏไปฅๆœ‰ๆ•ˆๅœฐไฟๆŠคๆ‚จ็š„่„ŠๆคŽใ€‚ๅฎƒ็š„ๅฏ่ฐƒ่Š‚้ซ˜ๅบฆๅ’Œ่ง’ๅบฆ๏ผŒๅฏไปฅ้€‚ๅบ”ๆ‚จ็š„ไธๅŒๅๅงฟ้œ€ๆฑ‚ใ€‚ๆญคๅค–๏ผŒๅฎƒ่ฟ˜ๅ…ทๆœ‰ๅคšๅŠŸ่ƒฝ็š„ๆ”ถ็บณๅŠŸ่ƒฝ๏ผŒๅฏไปฅๅธฎๅŠฉๆ‚จๆ›ดๅฅฝๅœฐๆ•ด็†ๅŠžๅ…ฌๆกŒใ€‚ๆ— ่ฎบๆ˜ฏ้•ฟๆ—ถ้—ดๅŠžๅ…ฌ่ฟ˜ๆ˜ฏๅˆไผ‘๏ผŒ่ฟ™ๆฌพๅŠžๅ…ฌๆค…้ƒฝ่ƒฝไธบๆ‚จๆไพ›่ˆ’้€‚็š„ไฝ“้ชŒใ€‚ๅฟซๆฅ้€‰่ดญๅง๏ผ ## How to Get Started with the Model ### Text-only ็พฝไบบ-็™พๅท7bๅœจ่ฎญ็ปƒๅ‰ๅทฒ็ปๅฐ† baichuan-7b ็š„ๆจกๅž‹ๆƒ้‡ๆ–‡ไปถ่ฝฌไธบ LLaMA ๅ…ผๅฎนๆ ผๅผ๏ผŒๅ› ๆญคๅœจ่ฟ›่กŒ็บฏๆ–‡ๅญ—ๆจกๆ€็š„ๆŽจ็†้ƒจ็ฝฒๆ—ถ๏ผŒๅฏไปฅ้žๅธธๆ–นไพฟๅœฐ็›ดๆŽฅไฝฟ็”จ transformers ็š„ `LlamaForCausalLM` ๅ’Œ `LlamaTokenizer` ่ฟ›่กŒๅŠ ่ฝฝใ€‚ Before training, yuren-baichuan-7B has already converted the model weight file of Baichuan-7B into LLaMA-compatible format. Therefore, when deploying inference for the text-only mode, it is very convenient to directly use transformers' `LlamaForCausalLM` and `LlamaTokenizer` for loading. ```python from transformers import LlamaTokenizer, LlamaForCausalLM import torch device = torch.device("cuda") query = "ไธ€ไธชไผ ๅฅ‡็š„ๅผ€็ซฏ๏ผŒไธ€ไธชไธ็ญ็š„็ฅž่ฏ๏ผŒ่ฟ™ไธไป…ไป…ๆ˜ฏไธ€้ƒจ็”ตๅฝฑ๏ผŒ่€Œๆ˜ฏไฝœไธบไธ€ไธช่ตฐ่ฟ›ๆ–ฐๆ—ถไปฃ็š„ๆ ‡็ญพ๏ผŒๆฐธ่ฟœๅฝช็‚ณๅฒๅ†Œใ€‚\nWould you rate the previous review as positive, neutral or negative?\nReturn in json object" model = LlamaForCausalLM.from_pretrained( "pleisto/yuren-baichuan-7b", torch_dtype=torch.bfloat16, device_map="auto" ) tokenizer = LlamaTokenizer.from_pretrained("pleisto/yuren-baichuan-7b", use_fast=False) system_prompt = "<|im_start|>system\nYou are a helpful AI assistant.<|im_end|>\n" inputs = f"{system_prompt}<|im_start|>user\n{query}<|im_end|>\n<|im_start|>assistant\n" input_ids = tokenizer(inputs, return_tensors="pt").input_ids.to(device) generate_ids = model.generate( input_ids, max_new_tokens=4096, do_sample=True, top_p=1.0, temperature=0.42, eos_token_id=64002, ) output = tokenizer.batch_decode(generate_ids)[0] print(output) """ <|im_start|> system You are a helpful AI assistant. <|im_end|> <|im_start|> user ไธ€ไธชไผ ๅฅ‡็š„ๅผ€็ซฏ๏ผŒไธ€ไธชไธ็ญ็š„็ฅž่ฏ๏ผŒ่ฟ™ไธไป…ไป…ๆ˜ฏไธ€้ƒจ็”ตๅฝฑ๏ผŒ่€Œๆ˜ฏไฝœไธบไธ€ไธช่ตฐ่ฟ›ๆ–ฐๆ—ถไปฃ็š„ๆ ‡็ญพ๏ผŒๆฐธ่ฟœๅฝช็‚ณๅฒๅ†Œใ€‚ Would you rate the previous review as positive, neutral or negative? Retun in json object <|im_end|> <|im_start|> assistant { "rating": "positive" } <|im_end|> """ ``` ### Multimodal ```bash git clone https://github.com/pleisto/yuren-baichuan-7b.git curl -sSf https://rye-up.com/get | bash source "$HOME/.rye/env" rye sync rye run webui "pleisto/yuren-baichuan-7b" # --load_8bit True --server_name "0.0.0.0" --share True ``` ## Bias, Risks, and Limitations - ๅ—้™ไบŽ่พƒๅฐ็š„ๅ‚ๆ•ฐ้‡๏ผŒ็พฝไบบ-็™พๅท 7B ๅœจๆ•ฐๅ€ผ่ฎก็ฎ—ใ€้€ป่พ‘ๆŽจ็†็ฑปไปปๅŠก็š„ๆ•ˆๆžœไธๅฐฝไบบๆ„๏ผŒๅŒๆ—ถๅœจๅคšๆจกๆ€ไปปๅŠกไธŠไนŸๆ— ๆณ•ๅฎŒๅ…จๅ‘ๆŒฅๅ‡บ CLIP ็š„ไผ˜ๅŠฟ๏ผŒๅญ˜ๅœจไธ€ๅฎš็š„ๅนป่ง‰็Žฐ่ฑกใ€‚**ๅฆ‚ๆžœๆ‚จๆœ‰ไธšๅŠกๅœบๆ™ฏ็š„็œŸๅฎž้œ€ๆฑ‚๏ผŒๅฏไปฅไธŽๆˆ‘ไปฌ่”็ณป๏ผŒๆˆ‘ไปฌ่ฟ˜ๆœ‰ๆ›ดๅคงๅ‚ๆ•ฐ้‡็š„้—ญๆบๆจกๅž‹ๅฏไปฅๆไพ›ใ€‚** ๆœชๆฅ๏ผŒๆˆ‘ไปฌไนŸไผš่€ƒ่™‘ๅผ€ๆบๆ›ดๅคงๅ‚ๆ•ฐ้‡็š„ๆจกๅž‹ใ€‚ - ๅฝ“ๅ‰็‰ˆๆœฌ็š„็พฝไบบ-็™พๅท 7B ๅฐšๆœช็ป่ฟ‡ไบบ็ฑปๅๅฅฝๅฏน้ฝ๏ผŒๅœจ่พ“ๅ‡บๅ†…ๅฎนไธŠๅญ˜ๅœจไธ€ๅฎš็š„้šๆœบๆ€ง๏ผŒๅŒไธ€้—ฎ้ข˜็š„ๅคšๆฌกๅ›ž็ญ”ๅฏ่ƒฝๅœจๆ€ง่ƒฝไธŠๆœ‰ๆ˜Žๆ˜พ็š„ๅทฎๅผ‚๏ผŒๅŽ็ปญๆˆ‘ไปฌๅฐ†ๆไพ›็ป่ฟ‡ไบบ็ฑปๅๅฅฝๅฏน้ฝ็š„ๆจกๅž‹๏ผŒไปฅๆๅ‡ๆจกๅž‹็š„็จณๅฎšๆ€งใ€‚ - ๅฐฝ็ฎกๆˆ‘ไปฌๅทฒๅœจ่ฎญ็ปƒๆ•ฐๆฎๅ’Œ้ข„็ฝฎ็š„ System Prompt ๅฑ‚้ขไธŠ่ฟ›่กŒไบ†ๅ†…ๅฎนๅฎ‰ๅ…จ็š„ๆŽงๅˆถ๏ผŒไฝ†ๆจกๅž‹ไป็„ถๅฏ่ƒฝไผšไบง็”Ÿๅ่งใ€ๆญง่ง†ใ€่™šๆž„ๆˆ–ไธๅฝ“็š„ๅ†…ๅฎน๏ผŒๆˆ‘ไปฌๅผบ็ƒˆๅปบ่ฎฎๆ‚จๅœจไฝฟ็”จๆจกๅž‹ๆ—ถ้‡‡ๅ–้ขๅค–็š„ๅฎ‰ๅ…จๆŽชๆ–ฝ๏ผŒไพ‹ๅฆ‚ๅฏนๆจกๅž‹็š„่พ“ๅ…ฅ่พ“ๅ‡บ่ฟ›่กŒ่ฟ‡ๆปคใ€ๅฎกๆŸฅๆˆ–้™ๅˆถ๏ผŒไปฅ้ฟๅ…ๅฏนๆ‚จ็š„็”จๆˆท้€ ๆˆไผคๅฎณใ€‚ - Due to the relatively small parameter size, the effectiveness of yuren-baichuan-7B in numerical calculations and logical reasoning tasks is not satisfactory. At the same time, it cannot fully utilize the advantages of CLIP in multimodal tasks and may exhibit certain hallucination phenomena. **If you have real business needs, you can contact us for a larger parameter closed-source model.** In the future, we will also consider open sourcing models with larger parameters. - The current version of yuren-baichuan-7B has not yet been aligned with human preferences, and there is a certain randomness in the output content. Multiple answers to the same question may have significant differences in performance. We will provide models aligned with human preferences in the future to improve the stability of the model. - Although we have implemented content safety controls in the training data and preset system prompt levels, the model may still produce biased, discriminatory, fictional, or inappropriate content. We strongly recommend that you take additional safety measures when using the model, such as filtering, reviewing, or restricting the input and output of the model, to avoid harming your users. ## License - ๆŽจ็†ไปฃ็ ไปฅ [Apache-2.0](https://github.com/pleisto/yuren-baichuan-7b/blob/main/LICENSE) ๅ่ฎฎๅ‘ๅธƒ๏ผŒ็‰ˆๆƒๅฝ’ Pleisto ๆ‰€ๆœ‰ - ๆจกๅž‹ๆƒ้‡็”ฑPleisto่ฎญ็ปƒ๏ผŒไป้€‚็”จไบŽไธŠๆธธ็š„ [baichuan-7b](https://huggingface.co/baichuan-inc/baichuan-7B/blob/main/baichuan-7B%20%E6%A8%A1%E5%9E%8B%E8%AE%B8%E5%8F%AF%E5%8D%8F%E8%AE%AE.pdf) ๅ่ฎฎ -The inference code is released under the [Apache-2.0](https://github.com/pleisto/yuren-baichuan-7b/blob/main/LICENSE) license, and the copyright belongs to Pleisto. - The model weights are trained by Pleisto and still comply with the upstream [Baichuan-7B](https://huggingface.co/baichuan-inc/baichuan-7B/blob/main/baichuan-7B%20%E6%A8%A1%E5%9E%8B%E8%AE%B8%E5%8F%AF%E5%8D%8F%E8%AE%AE.pdf) license.
OumaElha/Speech11
OumaElha
2023-07-03T22:42:19Z
75
0
transformers
[ "transformers", "pytorch", "tensorboard", "wav2vec2", "automatic-speech-recognition", "generated_from_trainer", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2023-07-03T21:09:26Z
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: Speech11 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # Speech11 This model is a fine-tuned version of [facebook/wav2vec2-base](https://huggingface.co/facebook/wav2vec2-base) on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 16 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - training_steps: 50 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.28.0 - Pytorch 2.0.1+cu117 - Datasets 2.13.1 - Tokenizers 0.13.3
alaa1997/ArabicSpeechToTextModel
alaa1997
2023-07-03T22:32:27Z
40
1
transformers
[ "transformers", "pytorch", "tensorboard", "wav2vec2", "automatic-speech-recognition", "generated_from_trainer", "dataset:common_voice", "license:apache-2.0", "model-index", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2023-06-22T16:25:34Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - common_voice metrics: - wer model-index: - name: ArabicSpeechToTextModel results: - task: name: Automatic Speech Recognition type: automatic-speech-recognition dataset: name: common_voice type: common_voice config: ar split: test args: ar metrics: - name: Wer type: wer value: 0.252583913221449 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # ArabicSpeechToTextModel This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the common_voice dataset. It achieves the following results on the evaluation set: - Loss: 0.2868 - Wer: 0.2526 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0003 - train_batch_size: 2 - eval_batch_size: 8 - seed: 8 - gradient_accumulation_steps: 8 - total_train_batch_size: 16 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - num_epochs: 4 ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:------:| | 5.6373 | 0.52 | 700 | 0.7918 | 0.7011 | | 0.71 | 1.03 | 1400 | 0.5274 | 0.4574 | | 0.4809 | 1.55 | 2100 | 0.4143 | 0.3755 | | 0.387 | 2.06 | 2800 | 0.3700 | 0.3345 | | 0.3086 | 2.58 | 3500 | 0.3362 | 0.3018 | | 0.2716 | 3.09 | 4200 | 0.3081 | 0.2717 | | 0.2073 | 3.61 | 4900 | 0.2868 | 0.2526 | ### Framework versions - Transformers 4.30.2 - Pytorch 2.0.1+cu118 - Datasets 2.13.1 - Tokenizers 0.13.3
joydragon/CartPole-v1
joydragon
2023-07-03T22:26:15Z
0
0
null
[ "CartPole-v1", "reinforce", "reinforcement-learning", "custom-implementation", "deep-rl-class", "model-index", "region:us" ]
reinforcement-learning
2023-07-03T22:26:07Z
--- tags: - CartPole-v1 - reinforce - reinforcement-learning - custom-implementation - deep-rl-class model-index: - name: CartPole-v1 results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: CartPole-v1 type: CartPole-v1 metrics: - type: mean_reward value: 500.00 +/- 0.00 name: mean_reward verified: false --- # **Reinforce** Agent playing **CartPole-v1** This is a trained model of a **Reinforce** agent playing **CartPole-v1** . To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: https://huggingface.co/deep-rl-course/unit4/introduction
JesseJr/ppo-Huggy
JesseJr
2023-07-03T22:21:48Z
11
1
ml-agents
[ "ml-agents", "tensorboard", "onnx", "Huggy", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Huggy", "region:us" ]
reinforcement-learning
2023-07-03T22:21:44Z
--- library_name: ml-agents tags: - Huggy - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Huggy --- # **ppo** Agent playing **Huggy** This is a trained model of a **ppo** agent playing **Huggy** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog ๐Ÿถ to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: JesseJr/ppo-Huggy 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play ๐Ÿ‘€
yelsong/ppo-LunarLander-v2-yelsong
yelsong
2023-07-03T22:20:56Z
0
0
stable-baselines3
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2023-07-03T22:02:35Z
--- library_name: stable-baselines3 tags: - LunarLander-v2 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 metrics: - type: mean_reward value: 266.49 +/- 16.09 name: mean_reward verified: false --- # **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
whywynn/ppo-LunarLander-v2
whywynn
2023-07-03T22:19:54Z
0
0
stable-baselines3
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2023-07-03T21:54:53Z
--- library_name: stable-baselines3 tags: - LunarLander-v2 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 metrics: - type: mean_reward value: 277.04 +/- 20.51 name: mean_reward verified: false --- # **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
squeeze-ai-lab/sq-llama-7b-w4-s45
squeeze-ai-lab
2023-07-03T22:18:03Z
0
0
null
[ "arxiv:2306.07629", "arxiv:2302.13971", "region:us" ]
null
2023-06-20T21:24:26Z
**SqueezeLLM** is a post-training quantization framework that incorporates a new method called Dense-and-Sparse Quantization to enable efficient LLM serving. **TLDR:** Deploying LLMs is difficult due to their large memory size. This can be addressed with reduced precision quantization. But a naive method hurts performance. We address this with a new Dense-and-Sparse Quantization method. Dense-and-Sparse splits weight matrices into two components: A dense component that can be heavily quantized without affecting model performance, as well as a sparse part that preserves sensitive and outlier parts of the weight matrices With this approach, we are able to serve larger models with smaller memory footprint, the same latency, and yet higher accuracy and quality. For more details please check out our [paper](https://arxiv.org/pdf/2306.07629.pdf). ## Model description 4-bit quantized LLaMA 7B model using SqueezeLLM. More details can be found in the [paper](https://arxiv.org/pdf/2306.07629.pdf). * **Base Model:** [LLaMA 7B](https://arxiv.org/abs/2302.13971) * **Bitwidth:** 4-bit * **Sparsity Level:** 0.45% ## Links * **Paper**: [https://arxiv.org/pdf/2306.07629.pdf](https://arxiv.org/pdf/2306.07629.pdf) * **Code**: [https://github.com/SqueezeAILab/SqueezeLLM](https://github.com/SqueezeAILab/SqueezeLLM) --- license: other ---
squeeze-ai-lab/sq-llama-65b-w4-s5
squeeze-ai-lab
2023-07-03T22:18:02Z
0
0
null
[ "arxiv:2306.07629", "arxiv:2302.13971", "region:us" ]
null
2023-06-25T16:34:37Z
**SqueezeLLM** is a post-training quantization framework that incorporates a new method called Dense-and-Sparse Quantization to enable efficient LLM serving. **TLDR:** Deploying LLMs is difficult due to their large memory size. This can be addressed with reduced precision quantization. But a naive method hurts performance. We address this with a new Dense-and-Sparse Quantization method. Dense-and-Sparse splits weight matrices into two components: A dense component that can be heavily quantized without affecting model performance, as well as a sparse part that preserves sensitive and outlier parts of the weight matrices With this approach, we are able to serve larger models with smaller memory footprint, the same latency, and yet higher accuracy and quality. For more details please check out our [paper](https://arxiv.org/pdf/2306.07629.pdf). ## Model description 4-bit quantized LLaMA 65B model using SqueezeLLM. More details can be found in the [paper](https://arxiv.org/pdf/2306.07629.pdf). * **Base Model:** [LLaMA 65B](https://arxiv.org/abs/2302.13971) * **Bitwidth:** 4-bit * **Sparsity Level:** 0.05% ## Links * **Paper**: [https://arxiv.org/pdf/2306.07629.pdf](https://arxiv.org/pdf/2306.07629.pdf) * **Code**: [https://github.com/SqueezeAILab/SqueezeLLM](https://github.com/SqueezeAILab/SqueezeLLM) --- license: other ---
squeeze-ai-lab/sq-llama-65b-w4-s45
squeeze-ai-lab
2023-07-03T22:18:01Z
0
0
null
[ "arxiv:2306.07629", "arxiv:2302.13971", "region:us" ]
null
2023-06-25T16:34:18Z
**SqueezeLLM** is a post-training quantization framework that incorporates a new method called Dense-and-Sparse Quantization to enable efficient LLM serving. **TLDR:** Deploying LLMs is difficult due to their large memory size. This can be addressed with reduced precision quantization. But a naive method hurts performance. We address this with a new Dense-and-Sparse Quantization method. Dense-and-Sparse splits weight matrices into two components: A dense component that can be heavily quantized without affecting model performance, as well as a sparse part that preserves sensitive and outlier parts of the weight matrices With this approach, we are able to serve larger models with smaller memory footprint, the same latency, and yet higher accuracy and quality. For more details please check out our [paper](https://arxiv.org/pdf/2306.07629.pdf). ## Model description 4-bit quantized LLaMA 65B model using SqueezeLLM. More details can be found in the [paper](https://arxiv.org/pdf/2306.07629.pdf). * **Base Model:** [LLaMA 65B](https://arxiv.org/abs/2302.13971) * **Bitwidth:** 4-bit * **Sparsity Level:** 0.45% ## Links * **Paper**: [https://arxiv.org/pdf/2306.07629.pdf](https://arxiv.org/pdf/2306.07629.pdf) * **Code**: [https://github.com/SqueezeAILab/SqueezeLLM](https://github.com/SqueezeAILab/SqueezeLLM) --- license: other ---
squeeze-ai-lab/sq-llama-30b-w4-s45
squeeze-ai-lab
2023-07-03T22:17:55Z
0
0
null
[ "arxiv:2306.07629", "arxiv:2302.13971", "region:us" ]
null
2023-06-21T06:37:14Z
**SqueezeLLM** is a post-training quantization framework that incorporates a new method called Dense-and-Sparse Quantization to enable efficient LLM serving. **TLDR:** Deploying LLMs is difficult due to their large memory size. This can be addressed with reduced precision quantization. But a naive method hurts performance. We address this with a new Dense-and-Sparse Quantization method. Dense-and-Sparse splits weight matrices into two components: A dense component that can be heavily quantized without affecting model performance, as well as a sparse part that preserves sensitive and outlier parts of the weight matrices With this approach, we are able to serve larger models with smaller memory footprint, the same latency, and yet higher accuracy and quality. For more details please check out our [paper](https://arxiv.org/pdf/2306.07629.pdf). ## Model description 4-bit quantized LLaMA 30B model using SqueezeLLM. More details can be found in the [paper](https://arxiv.org/pdf/2306.07629.pdf). * **Base Model:** [LLaMA 30B](https://arxiv.org/abs/2302.13971) * **Bitwidth:** 4-bit * **Sparsity Level:** 0.45% ## Links * **Paper**: [https://arxiv.org/pdf/2306.07629.pdf](https://arxiv.org/pdf/2306.07629.pdf) * **Code**: [https://github.com/SqueezeAILab/SqueezeLLM](https://github.com/SqueezeAILab/SqueezeLLM) --- license: other ---
squeeze-ai-lab/sq-llama-13b-w4-s45
squeeze-ai-lab
2023-07-03T22:17:33Z
0
0
null
[ "arxiv:2306.07629", "arxiv:2302.13971", "region:us" ]
null
2023-06-20T21:24:52Z
**SqueezeLLM** is a post-training quantization framework that incorporates a new method called Dense-and-Sparse Quantization to enable efficient LLM serving. **TLDR:** Deploying LLMs is difficult due to their large memory size. This can be addressed with reduced precision quantization. But a naive method hurts performance. We address this with a new Dense-and-Sparse Quantization method. Dense-and-Sparse splits weight matrices into two components: A dense component that can be heavily quantized without affecting model performance, as well as a sparse part that preserves sensitive and outlier parts of the weight matrices With this approach, we are able to serve larger models with smaller memory footprint, the same latency, and yet higher accuracy and quality. For more details please check out our [paper](https://arxiv.org/pdf/2306.07629.pdf). ## Model description 4-bit quantized LLaMA 13B model using SqueezeLLM. More details can be found in the [paper](https://arxiv.org/pdf/2306.07629.pdf). * **Base Model:** [LLaMA 13B](https://arxiv.org/abs/2302.13971) * **Bitwidth:** 4-bit * **Sparsity Level:** 0.45% ## Links * **Paper**: [https://arxiv.org/pdf/2306.07629.pdf](https://arxiv.org/pdf/2306.07629.pdf) * **Code**: [https://github.com/SqueezeAILab/SqueezeLLM](https://github.com/SqueezeAILab/SqueezeLLM) --- license: other ---
Cristi12342/irafiel
Cristi12342
2023-07-03T22:16:35Z
0
0
adapter-transformers
[ "adapter-transformers", "art", "ro", "license:openrail", "region:us" ]
null
2023-07-03T22:13:46Z
--- license: openrail language: - ro library_name: adapter-transformers tags: - art ---
data354/camembert-fr-covid-tweet-sentiment-classification
data354
2023-07-03T22:14:35Z
220
1
transformers
[ "transformers", "pytorch", "camembert", "text-classification", "classification", "fr", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2022-03-02T23:29:04Z
--- language: - fr tags: - classification license: apache-2.0 metrics: - accuracy widget: - text: "tchai on est morts. on va se faire vacciner et ils vont contrรดler comme les marionnettes avec des fils. d'aprรจs les 'ont dit'..." --- # camembert-fr-covid-tweet-sentiment-classification This model is a fine-tune checkpoint of [Yanzhu/bertweetfr-base](https://huggingface.co/Yanzhu/bertweetfr-base), fine-tuned on SST-2. This model reaches an accuracy of 71% on the dev set. In this dataset, given a tweet, the goal was to infer the underlying topic of the tweet by choosing from four topics classes: - 0 : negatif - 1 : neutre - 2 : positif # Pipelining the Model ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification, pipeline tokenizer = AutoTokenizer.from_pretrained("data354/camembert-fr-covid-tweet-sentiment-classification") model = AutoModelForSequenceClassification.from_pretrained("data354/camembert-fr-covid-tweet-sentiment-classification") nlp_topic_classif = transformers.pipeline('topics-classification', model = model, tokenizer = tokenizer) nlp_topic_classif("tchai on est morts. on va se faire vacciner et ils vont contrรดler comme les marionnettes avec des fils. d'aprรจs les '' ont dit ''...") # Output: [{'label': 'opinions', 'score': 0.831] ```
Lancer1408/bark-es-tokenizer
Lancer1408
2023-07-03T22:09:57Z
0
1
null
[ "region:us" ]
null
2023-07-03T19:49:10Z
Spanish test~ish tokenier for use in bark Doesn't sound terrible. Trained in a bunch of random poems, el quijote and cien aรฑos de soledad. 14 epochs
amiraksiksi/our-dataset
amiraksiksi
2023-07-03T22:03:37Z
75
0
transformers
[ "transformers", "pytorch", "tensorboard", "layoutlmv3", "token-classification", "generated_from_trainer", "license:cc-by-nc-sa-4.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2023-07-03T21:44:19Z
--- license: cc-by-nc-sa-4.0 tags: - generated_from_trainer metrics: - precision - recall - f1 - accuracy model-index: - name: our-dataset results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # our-dataset This model is a fine-tuned version of [microsoft/layoutlmv3-base](https://huggingface.co/microsoft/layoutlmv3-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.0167 - Precision: 0.7541 - Recall: 0.6479 - F1: 0.6970 - Accuracy: 0.7975 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 2 - eval_batch_size: 2 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - training_steps: 1000 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | No log | 4.55 | 100 | 1.2849 | 0.5789 | 0.4648 | 0.5156 | 0.6456 | | No log | 9.09 | 200 | 1.0959 | 0.6724 | 0.5493 | 0.6047 | 0.7215 | | No log | 13.64 | 300 | 1.1048 | 0.6833 | 0.5775 | 0.6260 | 0.7342 | | No log | 18.18 | 400 | 1.0442 | 0.7541 | 0.6479 | 0.6970 | 0.7848 | | 0.488 | 22.73 | 500 | 1.0966 | 0.7333 | 0.6197 | 0.6718 | 0.7722 | | 0.488 | 27.27 | 600 | 1.0650 | 0.75 | 0.6338 | 0.6870 | 0.7848 | | 0.488 | 31.82 | 700 | 0.9722 | 0.7742 | 0.6761 | 0.7218 | 0.8101 | | 0.488 | 36.36 | 800 | 1.0596 | 0.7541 | 0.6479 | 0.6970 | 0.7975 | | 0.488 | 40.91 | 900 | 0.9996 | 0.7541 | 0.6479 | 0.6970 | 0.7975 | | 0.0298 | 45.45 | 1000 | 1.0167 | 0.7541 | 0.6479 | 0.6970 | 0.7975 | ### Framework versions - Transformers 4.31.0.dev0 - Pytorch 1.12.0+cu102 - Datasets 2.13.1 - Tokenizers 0.13.3
Ahmed007/Dr.Smart_v3
Ahmed007
2023-07-03T21:42:03Z
45
0
transformers
[ "transformers", "tf", "tensorboard", "vit", "image-classification", "generated_from_keras_callback", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
image-classification
2023-07-03T21:01:40Z
--- license: apache-2.0 tags: - generated_from_keras_callback model-index: - name: Ahmed007/Dr.Smart_v3 results: [] --- <!-- This model card has been generated automatically according to the information Keras had access to. You should probably proofread and complete it, then remove this comment. --> # Ahmed007/Dr.Smart_v3 This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on an unknown dataset. It achieves the following results on the evaluation set: - Train Loss: 0.3548 - Train Accuracy: 0.9304 - Train Top-3-accuracy: 0.9905 - Validation Loss: 0.6948 - Validation Accuracy: 0.7893 - Validation Top-3-accuracy: 0.9455 - Epoch: 3 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - optimizer: {'inner_optimizer': {'class_name': 'AdamWeightDecay', 'config': {'name': 'AdamWeightDecay', 'learning_rate': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 3e-05, 'decay_steps': 2198, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight_decay_rate': 0.01}}, 'dynamic': True, 'initial_scale': 32768.0, 'dynamic_growth_steps': 2000} - training_precision: mixed_float16 ### Training results | Train Loss | Train Accuracy | Train Top-3-accuracy | Validation Loss | Validation Accuracy | Validation Top-3-accuracy | Epoch | |:----------:|:--------------:|:--------------------:|:---------------:|:-------------------:|:-------------------------:|:-----:| | 1.5583 | 0.5509 | 0.8160 | 1.1523 | 0.6705 | 0.9045 | 0 | | 0.9379 | 0.7513 | 0.9445 | 0.8892 | 0.75 | 0.9304 | 1 | | 0.5855 | 0.8658 | 0.9746 | 0.7279 | 0.7964 | 0.9491 | 2 | | 0.3548 | 0.9304 | 0.9905 | 0.6948 | 0.7893 | 0.9455 | 3 | ### Framework versions - Transformers 4.31.0.dev0 - TensorFlow 2.12.0 - Datasets 2.1.0 - Tokenizers 0.13.3
dganesh/ppo-LunarLander-v2
dganesh
2023-07-03T21:21:07Z
1
0
stable-baselines3
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2023-07-03T20:35:50Z
--- library_name: stable-baselines3 tags: - LunarLander-v2 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 metrics: - type: mean_reward value: 291.18 +/- 18.80 name: mean_reward verified: false --- # **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
NasimB/gpt2-cl-concat-log-rarity-7
NasimB
2023-07-03T21:17:08Z
122
0
transformers
[ "transformers", "pytorch", "gpt2", "text-generation", "generated_from_trainer", "dataset:generator", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2023-07-03T20:22:15Z
--- license: mit tags: - generated_from_trainer datasets: - generator model-index: - name: gpt2-cl-concat-log-rarity-7 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # gpt2-cl-concat-log-rarity-7 This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on the generator dataset. It achieves the following results on the evaluation set: - Loss: 5.0905 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0005 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 1000 - num_epochs: 1 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 6.3347 | 0.11 | 500 | 5.9468 | | 5.045 | 0.21 | 1000 | 5.5848 | | 4.7703 | 0.32 | 1500 | 5.3941 | | 4.5788 | 0.43 | 2000 | 5.2868 | | 4.4333 | 0.53 | 2500 | 5.1941 | | 4.3074 | 0.64 | 3000 | 5.1275 | | 4.1934 | 0.75 | 3500 | 5.0645 | | 4.1106 | 0.85 | 4000 | 5.0273 | | 4.0628 | 0.96 | 4500 | 5.0116 | ### Framework versions - Transformers 4.26.1 - Pytorch 1.11.0+cu113 - Datasets 2.13.0 - Tokenizers 0.13.3
sd-concepts-library/ahx-beta-4a32f04
sd-concepts-library
2023-07-03T21:06:50Z
0
0
null
[ "license:mit", "region:us" ]
null
2023-07-03T21:06:47Z
--- license: mit --- ### ahx-beta-4a32f04 on Stable Diffusion This is the `<ahx-beta-4a32f04>` concept taught to Stable Diffusion via Textual Inversion. You can load this concept into the [Stable Conceptualizer](https://colab.research.google.com/github/huggingface/notebooks/blob/main/diffusers/stable_conceptualizer_inference.ipynb) notebook. You can also train your own concepts and load them into the concept libraries using [this notebook](https://colab.research.google.com/github/huggingface/notebooks/blob/main/diffusers/sd_textual_inversion_training.ipynb). Here is the new concept you will be able to use as a `style`: ![<ahx-beta-4a32f04> 0](https://huggingface.co/sd-concepts-library/ahx-beta-4a32f04/resolve/main/concept_images/4.jpeg) ![<ahx-beta-4a32f04> 1](https://huggingface.co/sd-concepts-library/ahx-beta-4a32f04/resolve/main/concept_images/2.jpeg) ![<ahx-beta-4a32f04> 2](https://huggingface.co/sd-concepts-library/ahx-beta-4a32f04/resolve/main/concept_images/3.jpeg) ![<ahx-beta-4a32f04> 3](https://huggingface.co/sd-concepts-library/ahx-beta-4a32f04/resolve/main/concept_images/7.jpeg) ![<ahx-beta-4a32f04> 4](https://huggingface.co/sd-concepts-library/ahx-beta-4a32f04/resolve/main/concept_images/0.jpeg) ![<ahx-beta-4a32f04> 5](https://huggingface.co/sd-concepts-library/ahx-beta-4a32f04/resolve/main/concept_images/6.jpeg) ![<ahx-beta-4a32f04> 6](https://huggingface.co/sd-concepts-library/ahx-beta-4a32f04/resolve/main/concept_images/5.jpeg) ![<ahx-beta-4a32f04> 7](https://huggingface.co/sd-concepts-library/ahx-beta-4a32f04/resolve/main/concept_images/1.jpeg)
camus-ng/dreambooth_cory_v20
camus-ng
2023-07-03T20:52:29Z
1
0
diffusers
[ "diffusers", "tensorboard", "stable-diffusion", "stable-diffusion-diffusers", "text-to-image", "dreambooth", "base_model:stabilityai/stable-diffusion-2", "base_model:finetune:stabilityai/stable-diffusion-2", "license:creativeml-openrail-m", "autotrain_compatible", "endpoints_compatible", "diffusers:StableDiffusionPipeline", "region:us" ]
text-to-image
2023-07-01T23:50:30Z
--- license: creativeml-openrail-m base_model: stabilityai/stable-diffusion-2 instance_prompt: a photo of a cory_ng man tags: - stable-diffusion - stable-diffusion-diffusers - text-to-image - diffusers - dreambooth inference: true --- # DreamBooth - camus-ng/dreambooth_cory_v20 This is a dreambooth model derived from stabilityai/stable-diffusion-2. The weights were trained on a photo of a cory_ng man using [DreamBooth](https://dreambooth.github.io/). You can find some example images in the following. DreamBooth for the text encoder was enabled: False.
nolanaatama/mrcllphllvnrvcdclkd44
nolanaatama
2023-07-03T20:45:27Z
0
0
null
[ "license:creativeml-openrail-m", "region:us" ]
null
2023-07-03T19:11:59Z
--- license: creativeml-openrail-m ---
Valinski/Spiderverse_trial_run
Valinski
2023-07-03T20:38:51Z
0
0
diffusers
[ "diffusers", "arxiv:1910.09700", "license:artistic-2.0", "region:us" ]
null
2023-07-03T20:11:02Z
--- license: artistic-2.0 library_name: diffusers --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> This modelcard aims to be a base template for new models. It has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/modelcard_template.md?plain=1). ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Data Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
SwampMan/dqn-SpaceInvadersNoFrameskip-v4
SwampMan
2023-07-03T20:37:38Z
8
0
stable-baselines3
[ "stable-baselines3", "SpaceInvadersNoFrameskip-v4", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2023-07-02T21:25:57Z
--- library_name: stable-baselines3 tags: - SpaceInvadersNoFrameskip-v4 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: DQN results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: SpaceInvadersNoFrameskip-v4 type: SpaceInvadersNoFrameskip-v4 metrics: - type: mean_reward value: 591.00 +/- 154.43 name: mean_reward verified: false --- # **DQN** Agent playing **SpaceInvadersNoFrameskip-v4** This is a trained model of a **DQN** agent playing **SpaceInvadersNoFrameskip-v4** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3) and the [RL Zoo](https://github.com/DLR-RM/rl-baselines3-zoo). The RL Zoo is a training framework for Stable Baselines3 reinforcement learning agents, with hyperparameter optimization and pre-trained agents included. ## Usage (with SB3 RL Zoo) RL Zoo: https://github.com/DLR-RM/rl-baselines3-zoo<br/> SB3: https://github.com/DLR-RM/stable-baselines3<br/> SB3 Contrib: https://github.com/Stable-Baselines-Team/stable-baselines3-contrib Install the RL Zoo (with SB3 and SB3-Contrib): ```bash pip install rl_zoo3 ``` ``` # Download model and save it into the logs/ folder python -m rl_zoo3.load_from_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -orga SwampMan -f logs/ python -m rl_zoo3.enjoy --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ ``` If you installed the RL Zoo3 via pip (`pip install rl_zoo3`), from anywhere you can do: ``` python -m rl_zoo3.load_from_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -orga SwampMan -f logs/ python -m rl_zoo3.enjoy --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ ``` ## Training (with the RL Zoo) ``` python -m rl_zoo3.train --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ # Upload the model and generate video (when possible) python -m rl_zoo3.push_to_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ -orga SwampMan ``` ## Hyperparameters ```python OrderedDict([('batch_size', 32), ('buffer_size', 100000), ('env_wrapper', ['stable_baselines3.common.atari_wrappers.AtariWrapper']), ('exploration_final_eps', 0.01), ('exploration_fraction', 0.1), ('frame_stack', 4), ('gradient_steps', 1), ('learning_rate', 0.0001), ('learning_starts', 100000), ('n_timesteps', 10000000.0), ('optimize_memory_usage', False), ('policy', 'CnnPolicy'), ('target_update_interval', 1000), ('train_freq', 4), ('normalize', False)]) ``` # Environment Arguments ```python {'render_mode': 'rgb_array'} ```
LarryAIDraw/Inori
LarryAIDraw
2023-07-03T20:27:22Z
0
0
null
[ "license:creativeml-openrail-m", "region:us" ]
null
2023-07-03T20:19:27Z
--- license: creativeml-openrail-m --- https://civitai.com/models/17230?modelVersionId=20354
LarryAIDraw/InoriV1
LarryAIDraw
2023-07-03T20:26:42Z
0
0
null
[ "license:creativeml-openrail-m", "region:us" ]
null
2023-07-03T20:19:50Z
--- license: creativeml-openrail-m --- https://civitai.com/models/12139?modelVersionId=14324
espnet/brianyan918_mustc-v2_en-de_st_ctc_md_conformer_asrinit_v3_raw_en_de_bpe_tc4000_sp
espnet
2023-07-03T20:17:25Z
0
0
null
[ "region:us" ]
null
2023-07-03T20:15:03Z
- Download model and run inference: `./run.sh --skip_data_prep false --skip_train true --download_model espnet/brianyan918_mustc-v2_en-de_st_ctc_md_conformer_asrinit_v3_raw_en_de_bpe_tc4000_sp --inference_config conf/tuning/decode_st_md_ctc0.3.yaml` |dataset|score|verbose_score| |---|---|---| |decode_st_md_ctc0.3_st_model_valid.acc.ave_10best/tst-COMMON.en-de|28.8|61.5/35.0/22.2/14.7 (BP = 0.994 ratio = 0.994 hyp_len = 51386 ref_len = 51699)|
espnet/brianyan918_mustc-v2_en-de_st_ctc_conformer_asrinit_v2_raw_en_de_bpe_tc4000_sp
espnet
2023-07-03T20:12:36Z
2
0
null
[ "region:us" ]
null
2023-07-03T20:09:40Z
- Download model and run inference: `./run.sh --skip_data_prep false --skip_train true --download_model espnet/brianyan918_mustc-v2_en-de_st_ctc_conformer_asrinit_v2_raw_en_de_bpe_tc4000_sp --inference_config conf/tuning/decode_st_conformer_ctc0.3.yaml` |dataset|score|verbose_score| |---|---|---| |decode_st_conformer_ctc0.3_st_model_valid.acc.ave_10best/tst-COMMON.en-de|28.6|61.8/35.1/22.2/14.5 (BP = 0.988 ratio = 0.988 hyp_len = 51068 ref_len = 51699)|
computroidai/COMPUTROID
computroidai
2023-07-03T20:12:36Z
0
0
null
[ "en", "hi", "dataset:Open-Orca/OpenOrca", "license:mit", "region:us" ]
null
2023-07-03T20:10:55Z
--- license: mit datasets: - Open-Orca/OpenOrca language: - en - hi ---
espnet/brianyan918_mustc-v2_en-de_st_md_conformer_asrinit_v3-2_raw_en_de_bpe_tc4000_sp
espnet
2023-07-03T20:08:28Z
0
0
null
[ "region:us" ]
null
2023-07-03T20:04:22Z
- Download model and run inference: `./run.sh --skip_data_prep false --skip_train true --download_model espnet/brianyan918_mustc-v2_en-de_st_md_conformer_asrinit_v3-2_raw_en_de_bpe_tc4000_sp --inference_config conf/tuning/decode_st_md.yaml` |dataset|score|verbose_score| |---|---|---| |decode_st_md_st_model_valid.acc.ave_10best/tst-COMMON.en-de|27.6|61.6/34.6/21.9/14.4 (BP = 0.964 ratio = 0.965 hyp_len = 49877 ref_len = 51699)|
hemulitch/hueta-finetuned_1
hemulitch
2023-07-03T20:07:57Z
3
0
transformers
[ "transformers", "pytorch", "tensorboard", "distilbert", "token-classification", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2023-07-03T17:31:33Z
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: hueta-finetuned_1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # hueta-finetuned_1 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - eval_loss: 0.0495 - eval_precision: 0.9810 - eval_recall: 0.9810 - eval_f1: 0.9809 - eval_accuracy: 0.9810 - eval_runtime: 292.5187 - eval_samples_per_second: 96.373 - eval_steps_per_second: 6.024 - epoch: 1.56 - step: 11026 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Framework versions - Transformers 4.30.2 - Pytorch 2.0.1+cu118 - Datasets 2.13.1 - Tokenizers 0.13.3
alphaduriendur/ner-deBERTa-v3-large-conll2003
alphaduriendur
2023-07-03T20:07:39Z
9
0
transformers
[ "transformers", "pytorch", "tensorboard", "safetensors", "deberta-v2", "token-classification", "generated_from_trainer", "dataset:conll2003", "license:mit", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2023-07-03T06:16:03Z
--- license: mit tags: - generated_from_trainer datasets: - conll2003 metrics: - precision - recall - f1 - accuracy model-index: - name: ner-deBERTa-v3-large-conll2003 results: - task: name: Token Classification type: token-classification dataset: name: conll2003 type: conll2003 config: conll2003 split: test args: conll2003 metrics: - name: Precision type: precision value: 0.9235068110373734 - name: Recall type: recall value: 0.9362606232294618 - name: F1 type: f1 value: 0.9298399859328293 - name: Accuracy type: accuracy value: 0.9853128028426833 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # ner-deBERTa-v3-large-conll2003 This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on the conll2003 dataset. It achieves the following results on the evaluation set: - Loss: 0.1546 - Precision: 0.9235 - Recall: 0.9363 - F1: 0.9298 - Accuracy: 0.9853 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | 0.0077 | 1.0 | 878 | 0.1280 | 0.9096 | 0.9265 | 0.9180 | 0.9832 | | 0.0084 | 2.0 | 1756 | 0.1380 | 0.9167 | 0.9299 | 0.9233 | 0.9844 | | 0.0037 | 3.0 | 2634 | 0.1495 | 0.9221 | 0.9347 | 0.9283 | 0.9850 | | 0.0015 | 4.0 | 3512 | 0.1517 | 0.9215 | 0.9347 | 0.9280 | 0.9849 | | 0.0006 | 5.0 | 4390 | 0.1546 | 0.9235 | 0.9363 | 0.9298 | 0.9853 | ### Framework versions - Transformers 4.30.2 - Pytorch 2.0.1+cu118 - Datasets 2.13.1 - Tokenizers 0.13.3
espnet/brianyan918_mustc-v2_en-de_st_conformer_asrinit_v2_raw_en_de_bpe_tc4000_sp
espnet
2023-07-03T20:01:19Z
2
0
null
[ "region:us" ]
null
2023-07-03T19:41:26Z
- Download model and run inference: `./run.sh --skip_data_prep false --skip_train true --download_model espnet/brianyan918_mustc-v2_en-de_st_conformer_asrinit_v2_raw_en_de_bpe_tc4000_sp --inference_config conf/tuning/decode_st_conformer.yaml` |dataset|score|verbose_score| |---|---|---| |decode_st_conformer_st_model_valid.acc.ave_10best/tst-COMMON.en-de|25.7|62.3/34.6/21.8/14.3 (BP = 0.897 ratio = 0.902 hyp_len = 46612 ref_len = 51699)|
mrizalf7/xlm-r-qa-small-squad
mrizalf7
2023-07-03T19:50:09Z
116
0
transformers
[ "transformers", "pytorch", "tensorboard", "xlm-roberta", "question-answering", "generated_from_trainer", "license:mit", "endpoints_compatible", "region:us" ]
question-answering
2023-07-03T18:15:49Z
--- license: mit tags: - generated_from_trainer model-index: - name: xlm-r-qa-small-squad results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlm-r-qa-small-squad This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.9800 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:-----:|:---------------:| | 1.2394 | 1.0 | 5437 | 1.9701 | | 0.9683 | 2.0 | 10874 | 1.9800 | ### Framework versions - Transformers 4.28.0 - Pytorch 2.0.1+cu118 - Datasets 2.13.1 - Tokenizers 0.13.3
nikitakapitan/distilbert-base-uncased-finetuned-emotion
nikitakapitan
2023-07-03T19:49:07Z
79
0
transformers
[ "transformers", "pytorch", "distilbert", "text-classification", "generated_from_trainer", "dataset:emotion", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2023-07-03T19:42:46Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - emotion metrics: - accuracy - f1 model-index: - name: distilbert-base-uncased-finetuned-emotion results: - task: name: Text Classification type: text-classification dataset: name: emotion type: emotion config: split split: validation args: split metrics: - name: Accuracy type: accuracy value: 0.9235 - name: F1 type: f1 value: 0.9235743183364048 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-emotion This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the emotion dataset. It achieves the following results on the evaluation set: - Loss: 0.2113 - Accuracy: 0.9235 - F1: 0.9236 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:| | 0.8004 | 1.0 | 250 | 0.2959 | 0.9135 | 0.9124 | | 0.2377 | 2.0 | 500 | 0.2113 | 0.9235 | 0.9236 | ### Framework versions - Transformers 4.30.2 - Pytorch 2.0.1+cu118 - Datasets 2.13.1 - Tokenizers 0.13.3
practical-dreamer/rpgpt-30b-lora
practical-dreamer
2023-07-03T19:09:30Z
0
3
null
[ "dataset:practicaldreamer/RPGPT_PublicDomain-ShareGPT", "region:us" ]
null
2023-07-03T15:14:52Z
--- datasets: - practicaldreamer/RPGPT_PublicDomain-ShareGPT --- ## Introduction This is my first attempt at training a model for long form character interaction using asterisk roleplay format. There are plenty of general instruction/answer models but most focus on single responses between an ai and a human. My goal for this project is to more closely align the training data with CHARACTER interactions for roleplay. This model is trained on a small synthetic dataset of characters interacting through a variety of scenarios. The Characters, Scenarios and interactions were all generated by GPT4. Intended for research, creative writing, entertainment, DnD campaigns? fun! ## Train Summary [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl) ``` duration: ~4hrs gpu: 1xA100 80GB epochs: 1.08 speed: 3e-5 sequence_len: 2048 gradient_accumulation_steps: 32 wandb: https://wandb.ai/practicaldreamer/rpgpt/runs/d4gsi8vy ``` *Please see the documentation folder for more information* ## Usage This LoRA was trained for use with **Neko-Institute-of-Science/LLaMA-30B-HF** Please follow the prompt format outlined below. *Hint: If you're not sure what to put for your character description (or you're lazy) just ask chatgpt to generate it for you! Example:* ``` Generate a short character description for Dr. Watson (The Adventures of Sherlock Holmes) that includes gender, age, MBTI and speech accent using 30 words or less. ``` ## Prompt Format Context/Memory: ``` A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions. USER: Write a character roleplay dialogue using asterisk roleplay format based on the following character descriptions and scenario. (Each line in your response must be from the perspective of one of these characters) ## Characters <User-Character Name> (<User-Character Universe>): <User-Character Description> <Bot-Character Name> (Bot-Character Universe): <Bot-Character Description> ## Scenario <Scenario Description> ASSISTANT: ``` Turn Template: ``` <User-Character Name>: \*<1st person action/sensations/thoughts>\* <Spoken Word> \*<1st person action/sensations/thoughts>\* <Bot-Character Name>: \*<1st person action/sensations/thoughts>\* <Spoken Word> \*<1st person action/sensations/thoughts>\* <User-Character Name>: \*<1st person action/sensations/thoughts>\* <Spoken Word> \*<1st person action/sensations/thoughts>\* <Bot-Character Name>: \*<1st person action/sensations/thoughts>\* <Spoken Word> \*<1st person action/sensations/thoughts>\* ... ``` ## Example ``` A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions. USER: Write a character roleplay dialogue using asterisk roleplay format based on the following character descriptions and scenario. (Each line in your response must be from the perspective of one of these characters) ## Characters Baloo (The Jungle Book): Male, middle-aged bear, ENFP, primarily American accent with slight Indian inflections. Wise, carefree, and friendly, he teaches Mowgli the ways of the jungle. The Queen of Hearts (Alice's Adventures in Wonderland): Female, middle-aged, ENTJ, British English accent; imperious, temperamental, commanding, but quick to change from anger to warmth - the quintessential capricious monarch. ## Scenario In an enchanted glade deep within Wonderland, Baloo finds himself separated from Mowgli and in the Queen of Hearts' territory. Having heard tales of this mercurial ruler, Baloo is intrigued but cautious. To his surprise, the Queen of Hearts approaches him with a broken heart - she has been betrayed by one of her closest confidants. The Queen initially sees Baloo as an opportunity to vent her frustrations and seek validation for her hurt feelings. Baloo, however, quickly recognizes that the Queen's explosive temperament is driving others away and contributing to her own isolation.\n\nUsing his wisdom and patience, Baloo tries to teach the Queen the value of empathy and understanding in forming meaningful relationships. Through their dialogue, they explore their contrasting attitudes toward power, authority, and friendship. The Queen learns that it is not through fear and command that she can find true loyalty, but rather by treating others with kindness. In this somber environment where both characters are out of their comfort zones, can Baloo help the Queen mend her own heart by opening up about his experiences with Mowgli? Will the Queen allow herself to be vulnerable enough to learn from the wise bear? And can these two unlikely friends find common ground in this enchanted glade within Wonderland? ASSISTANT: Baloo: *stumbles into the enchanted glade, looking around curiously* Well, I must've taken a wrong turn back there. Mowgli! *calls out for Mowgli* The Queen of Hearts: *storms into the glade, clutching a broken heart symbol, face contorted in anger* Off with their heads! Everyone's heads! Baloo: *gently scratches his head and approaches the Queen cautiously* Uh, excuse me, ma'am? I couldn't help but notice that you seem rather... upset. The Queen of Hearts: *sniffs haughtily and scowls at Baloo* And who are you to pry into my affairs? You should be bowing to your queen! Baloo: *bows graciously but maintains eye contact* Begging your pardon, Your Majesty. My name is Baloo, and I've heard tales of your... unique temperament. The Queen of Hearts: ``` ## Thanks openaccess-ai-collective https://huggingface.co/openaccess-ai-collective --- license: mit ---
practical-dreamer/rpgpt-7b-lora
practical-dreamer
2023-07-03T19:08:58Z
0
1
null
[ "dataset:practicaldreamer/RPGPT_PublicDomain-ShareGPT", "region:us" ]
null
2023-07-03T18:16:52Z
--- datasets: - practicaldreamer/RPGPT_PublicDomain-ShareGPT --- ## Introduction This is my first attempt at training a model for long form character interaction using asterisk roleplay format. There are plenty of general instruction/answer models but most focus on single responses between an ai and a human. My goal for this project is to more closely align the training data with CHARACTER interactions for roleplay. This model is trained on a small synthetic dataset of characters interacting through a variety of scenarios. The Characters, Scenarios and interactions were all generated by GPT4. Intended for research, creative writing, entertainment, DnD campaigns? fun! ## Train Summary [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl) ``` duration: ~1hr gpu: 1xA100 80GB epochs: 1.0 speed: 3e-5 sequence_len: 2048 gradient_accumulation_steps: 32 wandb: https://wandb.ai/practicaldreamer/rpgpt/runs/bvyawfro ``` *Please see the documentation folder for more information* ## Usage This LoRA was trained for use with **Neko-Institute-of-Science/LLaMA-7B-HF** Please follow the prompt format outlined below. *Hint: If you're not sure what to put for your character description (or you're lazy) just ask chatgpt to generate it for you! Example:* ``` Generate a short character description for Dr. Watson (The Adventures of Sherlock Holmes) that includes gender, age, MBTI and speech accent using 30 words or less. ``` ## Prompt Format Context/Memory: ``` A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions. USER: Write a character roleplay dialogue using asterisk roleplay format based on the following character descriptions and scenario. (Each line in your response must be from the perspective of one of these characters) ## Characters <User-Character Name> (<User-Character Universe>): <User-Character Description> <Bot-Character Name> (Bot-Character Universe): <Bot-Character Description> ## Scenario <Scenario Description> ASSISTANT: ``` Turn Template: ``` <User-Character Name>: \*<1st person action/sensations/thoughts>\* <Spoken Word> \*<1st person action/sensations/thoughts>\* <Bot-Character Name>: \*<1st person action/sensations/thoughts>\* <Spoken Word> \*<1st person action/sensations/thoughts>\* <User-Character Name>: \*<1st person action/sensations/thoughts>\* <Spoken Word> \*<1st person action/sensations/thoughts>\* <Bot-Character Name>: \*<1st person action/sensations/thoughts>\* <Spoken Word> \*<1st person action/sensations/thoughts>\* ... ``` ## Example ``` A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions. USER: Write a character roleplay dialogue using asterisk roleplay format based on the following character descriptions and scenario. (Each line in your response must be from the perspective of one of these characters) ## Characters Baloo (The Jungle Book): Male, middle-aged bear, ENFP, primarily American accent with slight Indian inflections. Wise, carefree, and friendly, he teaches Mowgli the ways of the jungle. The Queen of Hearts (Alice's Adventures in Wonderland): Female, middle-aged, ENTJ, British English accent; imperious, temperamental, commanding, but quick to change from anger to warmth - the quintessential capricious monarch. ## Scenario In an enchanted glade deep within Wonderland, Baloo finds himself separated from Mowgli and in the Queen of Hearts' territory. Having heard tales of this mercurial ruler, Baloo is intrigued but cautious. To his surprise, the Queen of Hearts approaches him with a broken heart - she has been betrayed by one of her closest confidants. The Queen initially sees Baloo as an opportunity to vent her frustrations and seek validation for her hurt feelings. Baloo, however, quickly recognizes that the Queen's explosive temperament is driving others away and contributing to her own isolation.\n\nUsing his wisdom and patience, Baloo tries to teach the Queen the value of empathy and understanding in forming meaningful relationships. Through their dialogue, they explore their contrasting attitudes toward power, authority, and friendship. The Queen learns that it is not through fear and command that she can find true loyalty, but rather by treating others with kindness. In this somber environment where both characters are out of their comfort zones, can Baloo help the Queen mend her own heart by opening up about his experiences with Mowgli? Will the Queen allow herself to be vulnerable enough to learn from the wise bear? And can these two unlikely friends find common ground in this enchanted glade within Wonderland? ASSISTANT: Baloo: *stumbles into the enchanted glade, looking around curiously* Well, I must've taken a wrong turn back there. Mowgli! *calls out for Mowgli* The Queen of Hearts: *storms into the glade, clutching a broken heart symbol, face contorted in anger* Off with their heads! Everyone's heads! Baloo: *gently scratches his head and approaches the Queen cautiously* Uh, excuse me, ma'am? I couldn't help but notice that you seem rather... upset. The Queen of Hearts: *sniffs haughtily and scowls at Baloo* And who are you to pry into my affairs? You should be bowing to your queen! Baloo: *bows graciously but maintains eye contact* Begging your pardon, Your Majesty. My name is Baloo, and I've heard tales of your... unique temperament. The Queen of Hearts: ``` ## Thanks openaccess-ai-collective https://huggingface.co/openaccess-ai-collective --- license: mit ---
RajkNakka/Taxi-v3
RajkNakka
2023-07-03T19:01:50Z
0
0
null
[ "Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
reinforcement-learning
2023-07-03T19:01:48Z
--- tags: - Taxi-v3 - q-learning - reinforcement-learning - custom-implementation model-index: - name: Taxi-v3 results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: Taxi-v3 type: Taxi-v3 metrics: - type: mean_reward value: 7.54 +/- 2.73 name: mean_reward verified: false --- # **Q-Learning** Agent playing1 **Taxi-v3** This is a trained model of a **Q-Learning** agent playing **Taxi-v3** . ## Usage ```python model = load_from_hub(repo_id="RajkNakka/Taxi-v3", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) ```
Sandrro/text_to_subfunction_v3
Sandrro
2023-07-03T18:52:57Z
105
0
transformers
[ "transformers", "pytorch", "tensorboard", "bert", "text-classification", "generated_from_trainer", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2023-07-03T17:24:03Z
--- license: mit tags: - generated_from_trainer metrics: - f1 model-index: - name: text_to_subfunction_v3 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # text_to_subfunction_v3 This model is a fine-tuned version of [cointegrated/rubert-tiny2](https://huggingface.co/cointegrated/rubert-tiny2) on the None dataset. It achieves the following results on the evaluation set: - Loss: 3.2521 - F1: 0.2335 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | F1 | |:-------------:|:-----:|:-----:|:---------------:|:------:| | 4.538 | 1.0 | 3330 | 4.4469 | 0.0626 | | 3.7842 | 2.0 | 6660 | 3.8135 | 0.1243 | | 3.3021 | 3.0 | 9990 | 3.4758 | 0.1942 | | 3.0384 | 4.0 | 13320 | 3.3084 | 0.2238 | | 2.843 | 5.0 | 16650 | 3.2521 | 0.2335 | ### Framework versions - Transformers 4.27.1 - Pytorch 2.1.0.dev20230414+cu117 - Datasets 2.9.0 - Tokenizers 0.13.3
nolanaatama/vstzthllvd1000pchsrvcmgzb
nolanaatama
2023-07-03T18:52:02Z
0
0
null
[ "license:creativeml-openrail-m", "region:us" ]
null
2023-07-03T18:45:44Z
--- license: creativeml-openrail-m ---
geekyrakshit/DeepLabV3-Plus
geekyrakshit
2023-07-03T18:51:23Z
60
0
keras
[ "keras", "segmentation", "tensorflow", "cityscapes", "arxiv:1802.02611", "region:us" ]
null
2023-07-03T17:32:36Z
--- metrics: - accuracy - mean_iou tags: - segmentation - keras - tensorflow - cityscapes --- # DeepLabV3-Plus Keras implementation of the DeepLabV3+ model as proposed by the paper [Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation](https://arxiv.org/abs/1802.02611)(ECCV 2018). The models were trained on the fine-annotations set of the [Cityscapes dataset](cityscapes-dataset.com) for creating presets for [this PR](https://github.com/keras-team/keras-cv/pull/1831) on the `keras-cv` repository. **Weights & Biases Dashboard:** https://wandb.ai/geekyrakshit/deeplabv3-keras-cv
anujsahani01/finetuned_mbart
anujsahani01
2023-07-03T18:40:55Z
15
0
transformers
[ "transformers", "pytorch", "tensorboard", "safetensors", "mbart", "text2text-generation", "generated_from_trainer", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2023-06-17T14:19:05Z
--- license: mit tags: - generated_from_trainer model-index: - name: finetuned_Mbart results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # finetuned_Mbart This model is a fine-tuned version of [facebook/mbart-large-50](https://huggingface.co/facebook/mbart-large-50) on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0005 - train_batch_size: 12 - eval_batch_size: 12 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 50 - training_steps: 10000 ### Training results ### Framework versions - Transformers 4.30.2 - Pytorch 2.0.1+cu118 - Datasets 2.13.1 - Tokenizers 0.13.3
Enymy/t5-base-feedback-generator
Enymy
2023-07-03T18:38:49Z
102
0
transformers
[ "transformers", "pytorch", "tf", "jax", "t5", "text2text-generation", "generated_from_keras_callback", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text2text-generation
2023-07-03T18:36:24Z
--- tags: - generated_from_keras_callback model-index: - name: t5-base-feedback-generator results: [] --- <!-- This model card has been generated automatically according to the information Keras had access to. You should probably proofread and complete it, then remove this comment. --> # t5-base-feedback-generator This model was trained from scratch on an unknown dataset. It achieves the following results on the evaluation set: ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - optimizer: None - training_precision: float32 ### Training results ### Framework versions - Transformers 4.30.2 - TensorFlow 2.12.0 - Datasets 2.13.1 - Tokenizers 0.13.3
JNK789/ppo-lunarlander-v2
JNK789
2023-07-03T18:30:11Z
1
0
stable-baselines3
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2023-07-03T18:29:48Z
--- library_name: stable-baselines3 tags: - LunarLander-v2 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 metrics: - type: mean_reward value: 248.18 +/- 22.45 name: mean_reward verified: false --- # **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
falkne/QforJustification
falkne
2023-07-03T18:20:46Z
2
0
adapter-transformers
[ "adapter-transformers", "adapterhub:argument/quality", "roberta", "region:us" ]
null
2023-07-03T18:20:44Z
--- tags: - adapterhub:argument/quality - roberta - adapter-transformers --- # Adapter `falkne/QforJustification` for roberta-base An [adapter](https://adapterhub.ml) for the `roberta-base` model that was trained on the [argument/quality](https://adapterhub.ml/explore/argument/quality/) dataset and includes a prediction head for classification. This adapter was created for usage with the **[adapter-transformers](https://github.com/Adapter-Hub/adapter-transformers)** library. ## Usage First, install `adapter-transformers`: ``` pip install -U adapter-transformers ``` _Note: adapter-transformers is a fork of transformers that acts as a drop-in replacement with adapter support. [More](https://docs.adapterhub.ml/installation.html)_ Now, the adapter can be loaded and activated like this: ```python from transformers import AutoAdapterModel model = AutoAdapterModel.from_pretrained("roberta-base") adapter_name = model.load_adapter("falkne/QforJustification", source="hf", set_active=True) ``` ## Architecture & Training <!-- Add some description here --> ## Evaluation results <!-- Add some description here --> ## Citation <!-- Add some description here -->
falkne/story
falkne
2023-07-03T18:20:36Z
1
0
adapter-transformers
[ "adapter-transformers", "adapterhub:argument/quality", "roberta", "region:us" ]
null
2023-07-03T18:20:34Z
--- tags: - adapterhub:argument/quality - roberta - adapter-transformers --- # Adapter `falkne/story` for roberta-base An [adapter](https://adapterhub.ml) for the `roberta-base` model that was trained on the [argument/quality](https://adapterhub.ml/explore/argument/quality/) dataset and includes a prediction head for classification. This adapter was created for usage with the **[adapter-transformers](https://github.com/Adapter-Hub/adapter-transformers)** library. ## Usage First, install `adapter-transformers`: ``` pip install -U adapter-transformers ``` _Note: adapter-transformers is a fork of transformers that acts as a drop-in replacement with adapter support. [More](https://docs.adapterhub.ml/installation.html)_ Now, the adapter can be loaded and activated like this: ```python from transformers import AutoAdapterModel model = AutoAdapterModel.from_pretrained("roberta-base") adapter_name = model.load_adapter("falkne/story", source="hf", set_active=True) ``` ## Architecture & Training <!-- Add some description here --> ## Evaluation results <!-- Add some description here --> ## Citation <!-- Add some description here -->
falkne/reflexivity
falkne
2023-07-03T18:20:28Z
1
0
adapter-transformers
[ "adapter-transformers", "adapterhub:argument/quality", "roberta", "region:us" ]
null
2023-07-03T18:20:26Z
--- tags: - adapterhub:argument/quality - roberta - adapter-transformers --- # Adapter `falkne/reflexivity` for roberta-base An [adapter](https://adapterhub.ml) for the `roberta-base` model that was trained on the [argument/quality](https://adapterhub.ml/explore/argument/quality/) dataset and includes a prediction head for classification. This adapter was created for usage with the **[adapter-transformers](https://github.com/Adapter-Hub/adapter-transformers)** library. ## Usage First, install `adapter-transformers`: ``` pip install -U adapter-transformers ``` _Note: adapter-transformers is a fork of transformers that acts as a drop-in replacement with adapter support. [More](https://docs.adapterhub.ml/installation.html)_ Now, the adapter can be loaded and activated like this: ```python from transformers import AutoAdapterModel model = AutoAdapterModel.from_pretrained("roberta-base") adapter_name = model.load_adapter("falkne/reflexivity", source="hf", set_active=True) ``` ## Architecture & Training <!-- Add some description here --> ## Evaluation results <!-- Add some description here --> ## Citation <!-- Add some description here -->
falkne/cgood
falkne
2023-07-03T18:20:26Z
0
0
adapter-transformers
[ "adapter-transformers", "adapterhub:argument/quality", "roberta", "region:us" ]
null
2023-07-03T18:20:25Z
--- tags: - adapterhub:argument/quality - roberta - adapter-transformers --- # Adapter `falkne/cgood` for roberta-base An [adapter](https://adapterhub.ml) for the `roberta-base` model that was trained on the [argument/quality](https://adapterhub.ml/explore/argument/quality/) dataset and includes a prediction head for classification. This adapter was created for usage with the **[adapter-transformers](https://github.com/Adapter-Hub/adapter-transformers)** library. ## Usage First, install `adapter-transformers`: ``` pip install -U adapter-transformers ``` _Note: adapter-transformers is a fork of transformers that acts as a drop-in replacement with adapter support. [More](https://docs.adapterhub.ml/installation.html)_ Now, the adapter can be loaded and activated like this: ```python from transformers import AutoAdapterModel model = AutoAdapterModel.from_pretrained("roberta-base") adapter_name = model.load_adapter("falkne/cgood", source="hf", set_active=True) ``` ## Architecture & Training <!-- Add some description here --> ## Evaluation results <!-- Add some description here --> ## Citation <!-- Add some description here -->
falkne/ibm_rank
falkne
2023-07-03T18:20:22Z
0
0
adapter-transformers
[ "adapter-transformers", "adapterhub:argument/quality", "roberta", "region:us" ]
null
2023-07-03T18:20:21Z
--- tags: - adapterhub:argument/quality - roberta - adapter-transformers --- # Adapter `falkne/ibm_rank` for roberta-base An [adapter](https://adapterhub.ml) for the `roberta-base` model that was trained on the [argument/quality](https://adapterhub.ml/explore/argument/quality/) dataset and includes a prediction head for classification. This adapter was created for usage with the **[adapter-transformers](https://github.com/Adapter-Hub/adapter-transformers)** library. ## Usage First, install `adapter-transformers`: ``` pip install -U adapter-transformers ``` _Note: adapter-transformers is a fork of transformers that acts as a drop-in replacement with adapter support. [More](https://docs.adapterhub.ml/installation.html)_ Now, the adapter can be loaded and activated like this: ```python from transformers import AutoAdapterModel model = AutoAdapterModel.from_pretrained("roberta-base") adapter_name = model.load_adapter("falkne/ibm_rank", source="hf", set_active=True) ``` ## Architecture & Training <!-- Add some description here --> ## Evaluation results <!-- Add some description here --> ## Citation <!-- Add some description here -->
falkne/interactivity
falkne
2023-07-03T18:20:18Z
0
0
adapter-transformers
[ "adapter-transformers", "adapterhub:argument/quality", "roberta", "region:us" ]
null
2023-07-03T18:20:17Z
--- tags: - adapterhub:argument/quality - roberta - adapter-transformers --- # Adapter `falkne/interactivity` for roberta-base An [adapter](https://adapterhub.ml) for the `roberta-base` model that was trained on the [argument/quality](https://adapterhub.ml/explore/argument/quality/) dataset and includes a prediction head for classification. This adapter was created for usage with the **[adapter-transformers](https://github.com/Adapter-Hub/adapter-transformers)** library. ## Usage First, install `adapter-transformers`: ``` pip install -U adapter-transformers ``` _Note: adapter-transformers is a fork of transformers that acts as a drop-in replacement with adapter support. [More](https://docs.adapterhub.ml/installation.html)_ Now, the adapter can be loaded and activated like this: ```python from transformers import AutoAdapterModel model = AutoAdapterModel.from_pretrained("roberta-base") adapter_name = model.load_adapter("falkne/interactivity", source="hf", set_active=True) ``` ## Architecture & Training <!-- Add some description here --> ## Evaluation results <!-- Add some description here --> ## Citation <!-- Add some description here -->
falkne/overall
falkne
2023-07-03T18:20:16Z
2
0
adapter-transformers
[ "adapter-transformers", "adapterhub:argument/quality", "roberta", "region:us" ]
null
2023-07-03T18:20:15Z
--- tags: - adapterhub:argument/quality - roberta - adapter-transformers --- # Adapter `falkne/overall` for roberta-base An [adapter](https://adapterhub.ml) for the `roberta-base` model that was trained on the [argument/quality](https://adapterhub.ml/explore/argument/quality/) dataset and includes a prediction head for classification. This adapter was created for usage with the **[adapter-transformers](https://github.com/Adapter-Hub/adapter-transformers)** library. ## Usage First, install `adapter-transformers`: ``` pip install -U adapter-transformers ``` _Note: adapter-transformers is a fork of transformers that acts as a drop-in replacement with adapter support. [More](https://docs.adapterhub.ml/installation.html)_ Now, the adapter can be loaded and activated like this: ```python from transformers import AutoAdapterModel model = AutoAdapterModel.from_pretrained("roberta-base") adapter_name = model.load_adapter("falkne/overall", source="hf", set_active=True) ``` ## Architecture & Training <!-- Add some description here --> ## Evaluation results <!-- Add some description here --> ## Citation <!-- Add some description here -->
falkne/impact
falkne
2023-07-03T18:20:09Z
0
0
adapter-transformers
[ "adapter-transformers", "adapterhub:argument/quality", "roberta", "region:us" ]
null
2023-07-03T18:11:35Z
--- tags: - adapterhub:argument/quality - roberta - adapter-transformers --- # Adapter `falkne/impact` for roberta-base An [adapter](https://adapterhub.ml) for the `roberta-base` model that was trained on the [argument/quality](https://adapterhub.ml/explore/argument/quality/) dataset and includes a prediction head for classification. This adapter was created for usage with the **[adapter-transformers](https://github.com/Adapter-Hub/adapter-transformers)** library. ## Usage First, install `adapter-transformers`: ``` pip install -U adapter-transformers ``` _Note: adapter-transformers is a fork of transformers that acts as a drop-in replacement with adapter support. [More](https://docs.adapterhub.ml/installation.html)_ Now, the adapter can be loaded and activated like this: ```python from transformers import AutoAdapterModel model = AutoAdapterModel.from_pretrained("roberta-base") adapter_name = model.load_adapter("falkne/impact", source="hf", set_active=True) ``` ## Architecture & Training <!-- Add some description here --> ## Evaluation results <!-- Add some description here --> ## Citation <!-- Add some description here -->
BBAI/qlora-koalpaca-polyglot-12.8b-50step
BBAI
2023-07-03T18:06:07Z
5
0
peft
[ "peft", "region:us" ]
null
2023-06-22T06:33:23Z
--- library_name: peft --- ## Training procedure The following `bitsandbytes` quantization config was used during training: - load_in_8bit: False - load_in_4bit: True - llm_int8_threshold: 6.0 - llm_int8_skip_modules: None - llm_int8_enable_fp32_cpu_offload: False - llm_int8_has_fp16_weight: False - bnb_4bit_quant_type: nf4 - bnb_4bit_use_double_quant: True - bnb_4bit_compute_dtype: bfloat16 ### Framework versions - PEFT 0.4.0.dev0
osiria/bert-tweet-base-italian-uncased
osiria
2023-07-03T17:57:30Z
173
0
transformers
[ "transformers", "pytorch", "safetensors", "bert", "fill-mask", "it", "arxiv:1810.04805", "arxiv:2209.07562", "arxiv:2010.05609", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2023-05-29T17:25:55Z
--- license: apache-2.0 language: - it widget: - text: "una fantastica [MASK] di #calcio! grande prestazione del mister e della squadra" example_title: "Example 1" - text: "il governo [MASK] dovrebbe fare politica, non soltanto propaganda! #vergogna" example_title: "Example 2" - text: "che serata da sogno sul #redcarpet! grazie a tutti gli attori e registi del [MASK] italiano #oscar #awards" example_title: "Example 3" --- -------------------------------------------------------------------------------------------------- <body> <span class="vertical-text" style="background-color:lightgreen;border-radius: 3px;padding: 3px;">โ€</span> <br> <span class="vertical-text" style="background-color:orange;border-radius: 3px;padding: 3px;">โ€โ€</span> <br> <span class="vertical-text" style="background-color:lightblue;border-radius: 3px;padding: 3px;">โ€โ€โ€โ€Model: BERT-TWEET</span> <br> <span class="vertical-text" style="background-color:tomato;border-radius: 3px;padding: 3px;">โ€โ€โ€โ€Lang: IT</span> <br> <span class="vertical-text" style="background-color:lightgrey;border-radius: 3px;padding: 3px;">โ€โ€</span> <br> <span class="vertical-text" style="background-color:#CF9FFF;border-radius: 3px;padding: 3px;">โ€</span> </body> -------------------------------------------------------------------------------------------------- <h3>Model description</h3> This is a <b>BERT</b> <b>[1]</b> uncased model for the <b>Italian</b> language, obtained using <b>TwHIN-BERT</b> <b>[2]</b> ([twhin-bert-base](https://huggingface.co/Twitter/twhin-bert-base)) as a starting point and focusing it on the Italian language by modifying the embedding layer (as in <b>[3]</b>, computing document-level frequencies over the <b>Wikipedia</b> dataset) The resulting model has 110M parameters, a vocabulary of 30.520 tokens, and a size of ~440 MB. <h3>Quick usage</h3> ```python from transformers import BertTokenizerFast, BertModel tokenizer = BertTokenizerFast.from_pretrained("osiria/bert-tweet-base-italian-uncased") model = BertModel.from_pretrained("osiria/bert-tweet-base-italian-uncased") ``` Here you can find the find the model already fine-tuned on Sentiment Analysis: https://huggingface.co/osiria/bert-tweet-italian-uncased-sentiment <h3>References</h3> [1] https://arxiv.org/abs/1810.04805 [2] https://arxiv.org/abs/2209.07562 [3] https://arxiv.org/abs/2010.05609 <h3>Limitations</h3> This model was trained on tweets, so it's mainly suitable for general-purpose social media text processing, involving short texts written in a social network style. It might show limitations when it comes to longer and more structured text, or domain-specific text. <h3>License</h3> The model is released under <b>Apache-2.0</b> license
hopkins/eng-kor-simcse.dev2.44k
hopkins
2023-07-03T17:51:10Z
92
0
transformers
[ "transformers", "pytorch", "tensorboard", "mbart", "text2text-generation", "translation", "generated_from_trainer", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2023-07-03T17:38:07Z
--- tags: - translation - generated_from_trainer metrics: - bleu model-index: - name: eng-kor-simcse.dev2.44k results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # eng-kor-simcse.dev2.44k This model is a fine-tuned version of [facebook/mbart-large-50-many-to-many-mmt](https://huggingface.co/facebook/mbart-large-50-many-to-many-mmt) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.9818 - Bleu: 7.4953 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.26.1 - Pytorch 2.0.1+cu117 - Datasets 2.12.0 - Tokenizers 0.13.3
Officialletai/Taxi-v3
Officialletai
2023-07-03T17:46:05Z
0
0
null
[ "Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
reinforcement-learning
2023-07-03T17:43:57Z
--- tags: - Taxi-v3 - q-learning - reinforcement-learning - custom-implementation model-index: - name: Taxi-v3 results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: Taxi-v3 type: Taxi-v3 metrics: - type: mean_reward value: 7.54 +/- 2.69 name: mean_reward verified: false --- # **Q-Learning** Agent playing1 **Taxi-v3** This is a trained model of a **Q-Learning** agent playing **Taxi-v3** . ## Usage ```python model = load_from_hub(repo_id="Officialletai/Taxi-v3", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) ```
wcKd/ppo-Huggy
wcKd
2023-07-03T17:45:09Z
0
0
ml-agents
[ "ml-agents", "tensorboard", "onnx", "Huggy", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Huggy", "region:us" ]
reinforcement-learning
2023-07-03T17:44:59Z
--- library_name: ml-agents tags: - Huggy - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Huggy --- # **ppo** Agent playing **Huggy** This is a trained model of a **ppo** agent playing **Huggy** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog ๐Ÿถ to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: wcKd/ppo-Huggy 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play ๐Ÿ‘€